diff --git a/.gitattributes b/.gitattributes index 02bda845a69b2cf77dc268147ab1f1a89c5cfd6c..ff94f51519b0a0b78a58188f3d58993c0361da28 100644 --- a/.gitattributes +++ b/.gitattributes @@ -160,3 +160,93 @@ HVU_QA/t5-viet-qg-finetuned/checkpoint-26000/spiece.model filter=lfs diff=lfs me HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/model.safetensors filter=lfs diff=lfs merge=lfs -text HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/optimizer.pt filter=lfs diff=lfs merge=lfs -text HVU_QA/t5-viet-qg-finetuned/checkpoint-26500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/spiece.model filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/model.safetensors filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/optimizer.pt filter=lfs diff=lfs merge=lfs -text +HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/spiece.model filter=lfs diff=lfs merge=lfs -text diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3e27d538b45a4a5d996f815454faab8b977bb124 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cea2862c3a7cab28cd11a37bc8a7c8b2289262c9e5c40308e80409afbbe8e305 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..2125946260a7ed45af33160029b89b75567c5914 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:44586e948497eb4afba40991bb3da45055b1603f67d02b37e1299d1809c757ea +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..cdbf6836f6010ae72f56bf184a79fdce8ad8bd45 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..ded887f8cb8b61c29bf4ed36881b4edd744e98cb Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..9635d2c918bac7cbab6f34cc7af7b9635d2dc0c9 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/trainer_state.json @@ -0,0 +1,18934 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.4211642852686092, + "eval_steps": 500, + "global_step": 27000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.644186304512e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3d8c04ff624c436b684357819636dcae674d1e46 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5907f7682b70be714c9ec0efb6d0615b3daac7ec6e91051cf2e93bd8380d1879 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..970c95ee7351cf417ea8239ad15abee758a2af20 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc909a24cd4f00202e6c09ac8624d045024e3ac07112a186c59bb786e0170b2f +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..44785c6dd655d8585992f671362fe111590ce5ff Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..6efd4af02e5bd4bb019d3cf8a39eea44cf8988d2 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..a660512f257cce9304e3eb1815335fe5cf9530ed --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/trainer_state.json @@ -0,0 +1,19284 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.42896362388469456, + "eval_steps": 500, + "global_step": 27500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.67463419904e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-27500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f1d231fb4a5538f03fd1a5e79b06c13b3c6e5adb --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7dff81bb9ad91523e26a78296f8df5b80f544a88df3f5189b055328e2d7b80f1 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..29369ecf37d773f7c1dfb8eec44538d4167ca505 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ba7ed3c12a5be04120f4bc7acc560c8a081284f9d37c15ebe29c5eee8c82dd9 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..57e2410298e6bbb425e53e83b833f2e99441d191 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..d6a07628d7f15ae3150227a603ae47b82b3036ac Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..5278d61bf2bbdfcc1d94ef84e6de0e519a78d109 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/trainer_state.json @@ -0,0 +1,19634 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.43676296250077995, + "eval_steps": 500, + "global_step": 28000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.705082093568e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..491ebb2599aff662ec3bd7701dad13c56720c1ba --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:20327de8126aa8b5d343941cfde767ce4f3b565c7baf502c3c9e924b023e201c +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..5081cf2d3a4e2fbc82f080e9c74ae68e45fd7879 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:85e6f54847a1dbbef470a29e0757c0c1029ad8e79f7b70cbd752ddbf503a52e8 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..3c2e206dae590bb4444ab3d07d196f1e8432ce42 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..359f3890bef03c8a0e9a9f41401ecb97a859908f Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..c930efb11c1b8f60932f5cd1b7264fde122fa0e4 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/trainer_state.json @@ -0,0 +1,19984 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.4445623011168653, + "eval_steps": 500, + "global_step": 28500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.735529988096e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-28500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dc3fd04aabb866708437ba3a568d1449d3ff4d79 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:82098f27204155045cb603a928a5764ecbea092cbad4386e6d4c7f6ccb8ff910 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..b600f7bda7c2210cb2f811d5570626a28f505ed5 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7f9b850b7aff10189548a36b854983ae256ae6bbf3aa627ac2bd6417fd17c57 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..84d81ef42bb5877317274e159411185565727680 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..511825cb2406dde805f98d9ab84f556ce1d3f1d1 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..39248d55b7045485cb4b44a4cd8381710af7d01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/trainer_state.json @@ -0,0 +1,20334 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.4523616397329506, + "eval_steps": 500, + "global_step": 29000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.765977882624e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bd1140b7c8354c0537fa5a49c295a8f198a3cb19 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7da2754ab312cb3beab8af6c7db3958b13d139070300e913f850f423f3e6cd05 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..ccd43812df7c9c66159e3c8d37fb337e64d3d746 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b1783f3caf782060629c33111b7b7e976af69275636ead5f213591fa33381f87 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..355cf1720f737d96beeb86833f387dfffa98afef Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..2e30cfb613cb57610727e66d362f596d6af8d555 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..4f5017531456051cb9cdc3028e19cedda24a9011 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/trainer_state.json @@ -0,0 +1,20684 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.460160978349036, + "eval_steps": 500, + "global_step": 29500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.796425777152e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-29500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4803eb35b5abb79b49e71269e22f866ada12c635 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1279b78e16d34a8083af1cc1d109cdbdd162f450f5db5ce9628473d68158ec0d +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..6a76c46da56a666e5b26fbd2fddf98e1a413d468 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3149015dc920d4fd8236b3134c7bf19840804ab2879ebff53584f9d1d12e34b1 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..5a10765b96d4432cdaea4e9fbea941de35664399 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..efa25cdb5d9b956747861b4ad9e81194a8d451e4 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..c0169f2e6e1670bba183b341ba2bac7ed67460b4 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/trainer_state.json @@ -0,0 +1,21034 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.46796031696512136, + "eval_steps": 500, + "global_step": 30000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.82687367168e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9b19529706a2012b4530b4e451c00fe9c0a5603d --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:86dee78d5699f1ad7d52bb6af99e90decc3494500c994dead0e824885c9407b5 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..62cac14845efdd48f0e2051438d9246502090514 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58603d4e96b5de46133797bd2a4ec937d0cc2f3b512c617418617622720e338f +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..74077eb7873430a3689757999d975253c82e9c5d Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..c5cc416271d19050476c6a34995297637b09d504 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..3253a8d95210879a706a25fcc6ba699488d79a14 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/trainer_state.json @@ -0,0 +1,21384 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.4757596555812067, + "eval_steps": 500, + "global_step": 30500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.857321566208e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-30500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..72c66368dc3f5aea8648decf15beac6222b5031c --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2720db6b18138e1824e2bba41306d97b8848b1b48eb22dfcdee3f1138246a15e +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..636b80a8cc8202b6dc3088b26121c8816e70ecf3 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee0b5667ed1973bd1bd187d022a77ec06da9d7b4316670e7c7916617d2767dd9 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..e7285294013b82959c3f25bbc16fd4f7adcaa7e6 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..9416ec9d7c84b777414fe7a84e214bc741099d1c Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..37b7fcacc5beb02169dee96425befcfe15043597 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/trainer_state.json @@ -0,0 +1,21734 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.4835589941972921, + "eval_steps": 500, + "global_step": 31000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.887769460736e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c81dac767c89be9f44e5065285962e6bab81e681 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a9324bc448b7ff835977378ccbfc07093aafe86ed7d26fe2e34713e69449898f +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..73e2894a3d19c4d61c216eb17a7006e703e9b822 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd692053b395cb2345f3abf7487c7f3eb01132833e015c596077c77e00cc6e75 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..6d617c22ca8274691858d2751a547a7dabacf4cb Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..dfc2764a85473103cb8f28c6d3a950ec3b4debdd Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..0b46922647111a2b8616ba68ff26475d300a32dd --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/trainer_state.json @@ -0,0 +1,22084 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.49135833281337743, + "eval_steps": 500, + "global_step": 31500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.918217355264e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-31500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..81a55e998c3c12445f33d39043147c09f738ca74 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:741496a63554f9bd7a6f79caa9d09359d3a612b99501d472d988f35378ab3942 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..fa6844625e33449de3b5704fa61de5066c05dd8c --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:85520169c8991b84f6719a7f6f8395fdda31b5b25be0196d25db754766afdc77 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..c25204e80c0d5310a67215d3683b5588b31807cc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..6b6015126122e5743cc5154e0ef39e390b9226c1 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..43045a766ce1b4d73217484eb527fe0ba551e9c9 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/trainer_state.json @@ -0,0 +1,22434 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.49915767142946277, + "eval_steps": 500, + "global_step": 32000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.948665249792e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..249a293bc579083afce240ee14bcec2e689d59c0 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39907a8a1e34920be8f29f20bb1e7a19616113e58fd1a1db7c5509a3a80cb6dd +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..27057947945c37aa3a92ce6e19ec53e8f09c26ac --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:50f0bb2a08830f4a6206b75029994c48bbfb906ba955bf9b0a43becbf7f49ff2 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..ac5fb371ff392a8e74040af01e429b55a5d996b6 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..1b767c6707044c5958b8357ad7e31da8682bee8e Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..74e90224daff39794aa020a722794f3bba7f6df4 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/trainer_state.json @@ -0,0 +1,22784 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5069570100455482, + "eval_steps": 500, + "global_step": 32500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.97911314432e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-32500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..367dee4e345691d98467aa2c14eba17854b649a2 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:779d47169eb27077c207564b37b50a49b66026dd33f5e59927966defd34066e3 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..2224dc7d7756666f24a603b6c6077c28db73a73c --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb9b4b0e342d99b401abea7057c37da244680cd85e73513ddd50e7030d373a53 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..25d554de371bc95f3c49b411994c8f9aca7c179e Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..1216d9dc1bc52944dd7c5ce4a17d610348e76d4a Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..a24a9808628fc811ce37049ad063710e78bf312a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/trainer_state.json @@ -0,0 +1,23134 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5147563486616334, + "eval_steps": 500, + "global_step": 33000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + }, + { + "epoch": 0.5071129968178698, + "grad_norm": 1.0075199604034424, + "learning_rate": 0.00016619350679062414, + "loss": 0.2569, + "step": 32510 + }, + { + "epoch": 0.5072689835901916, + "grad_norm": 1.3489066362380981, + "learning_rate": 0.0001661831076724694, + "loss": 0.2565, + "step": 32520 + }, + { + "epoch": 0.5074249703625132, + "grad_norm": 0.2534058094024658, + "learning_rate": 0.00016617270855431458, + "loss": 0.2292, + "step": 32530 + }, + { + "epoch": 0.507580957134835, + "grad_norm": 0.31723281741142273, + "learning_rate": 0.00016616230943615983, + "loss": 0.1467, + "step": 32540 + }, + { + "epoch": 0.5077369439071566, + "grad_norm": 0.29659849405288696, + "learning_rate": 0.00016615191031800502, + "loss": 0.2768, + "step": 32550 + }, + { + "epoch": 0.5078929306794784, + "grad_norm": 1.5246435403823853, + "learning_rate": 0.00016614151119985027, + "loss": 0.216, + "step": 32560 + }, + { + "epoch": 0.5080489174518001, + "grad_norm": 0.6026854515075684, + "learning_rate": 0.00016613111208169546, + "loss": 0.2277, + "step": 32570 + }, + { + "epoch": 0.5082049042241218, + "grad_norm": 0.41994890570640564, + "learning_rate": 0.0001661207129635407, + "loss": 0.2433, + "step": 32580 + }, + { + "epoch": 0.5083608909964435, + "grad_norm": 2.084599733352661, + "learning_rate": 0.00016611031384538593, + "loss": 0.4081, + "step": 32590 + }, + { + "epoch": 0.5085168777687652, + "grad_norm": 0.2945229709148407, + "learning_rate": 0.00016609991472723115, + "loss": 0.1329, + "step": 32600 + }, + { + "epoch": 0.508672864541087, + "grad_norm": 1.4981423616409302, + "learning_rate": 0.00016608951560907636, + "loss": 0.2855, + "step": 32610 + }, + { + "epoch": 0.5088288513134086, + "grad_norm": 0.961737871170044, + "learning_rate": 0.00016607911649092158, + "loss": 0.1654, + "step": 32620 + }, + { + "epoch": 0.5089848380857304, + "grad_norm": 1.49734628200531, + "learning_rate": 0.0001660687173727668, + "loss": 0.2081, + "step": 32630 + }, + { + "epoch": 0.509140824858052, + "grad_norm": 2.538135051727295, + "learning_rate": 0.00016605831825461202, + "loss": 0.3023, + "step": 32640 + }, + { + "epoch": 0.5092968116303738, + "grad_norm": 0.2981272041797638, + "learning_rate": 0.00016604791913645724, + "loss": 0.214, + "step": 32650 + }, + { + "epoch": 0.5094527984026954, + "grad_norm": 1.8863650560379028, + "learning_rate": 0.00016603752001830246, + "loss": 0.3363, + "step": 32660 + }, + { + "epoch": 0.5096087851750172, + "grad_norm": 0.022720983251929283, + "learning_rate": 0.00016602712090014768, + "loss": 0.2463, + "step": 32670 + }, + { + "epoch": 0.5097647719473388, + "grad_norm": 1.154977798461914, + "learning_rate": 0.0001660167217819929, + "loss": 0.4484, + "step": 32680 + }, + { + "epoch": 0.5099207587196606, + "grad_norm": 0.9470499157905579, + "learning_rate": 0.00016600632266383812, + "loss": 0.1196, + "step": 32690 + }, + { + "epoch": 0.5100767454919822, + "grad_norm": 0.19955401122570038, + "learning_rate": 0.00016599592354568334, + "loss": 0.1762, + "step": 32700 + }, + { + "epoch": 0.510232732264304, + "grad_norm": 0.4098282754421234, + "learning_rate": 0.00016598552442752856, + "loss": 0.2739, + "step": 32710 + }, + { + "epoch": 0.5103887190366257, + "grad_norm": 0.21697258949279785, + "learning_rate": 0.00016597512530937378, + "loss": 0.2555, + "step": 32720 + }, + { + "epoch": 0.5105447058089474, + "grad_norm": 4.894479751586914, + "learning_rate": 0.000165964726191219, + "loss": 0.3137, + "step": 32730 + }, + { + "epoch": 0.5107006925812692, + "grad_norm": 0.8396711349487305, + "learning_rate": 0.00016595432707306422, + "loss": 0.3105, + "step": 32740 + }, + { + "epoch": 0.5108566793535908, + "grad_norm": 1.5494674444198608, + "learning_rate": 0.00016594392795490944, + "loss": 0.2271, + "step": 32750 + }, + { + "epoch": 0.5110126661259126, + "grad_norm": 1.3409888744354248, + "learning_rate": 0.00016593352883675466, + "loss": 0.1796, + "step": 32760 + }, + { + "epoch": 0.5111686528982342, + "grad_norm": 1.7590911388397217, + "learning_rate": 0.00016592312971859988, + "loss": 0.2185, + "step": 32770 + }, + { + "epoch": 0.511324639670556, + "grad_norm": 0.7350002527236938, + "learning_rate": 0.0001659127306004451, + "loss": 0.1911, + "step": 32780 + }, + { + "epoch": 0.5114806264428776, + "grad_norm": 0.7176411747932434, + "learning_rate": 0.00016590233148229032, + "loss": 0.6918, + "step": 32790 + }, + { + "epoch": 0.5116366132151994, + "grad_norm": 2.5005321502685547, + "learning_rate": 0.00016589193236413554, + "loss": 0.2329, + "step": 32800 + }, + { + "epoch": 0.511792599987521, + "grad_norm": 1.3916250467300415, + "learning_rate": 0.00016588153324598076, + "loss": 0.2102, + "step": 32810 + }, + { + "epoch": 0.5119485867598428, + "grad_norm": 0.721671998500824, + "learning_rate": 0.00016587113412782598, + "loss": 0.2965, + "step": 32820 + }, + { + "epoch": 0.5121045735321644, + "grad_norm": 0.6120114326477051, + "learning_rate": 0.0001658607350096712, + "loss": 0.2661, + "step": 32830 + }, + { + "epoch": 0.5122605603044862, + "grad_norm": 0.29123347997665405, + "learning_rate": 0.00016585033589151642, + "loss": 0.2455, + "step": 32840 + }, + { + "epoch": 0.5124165470768078, + "grad_norm": 2.7300868034362793, + "learning_rate": 0.00016583993677336164, + "loss": 0.2049, + "step": 32850 + }, + { + "epoch": 0.5125725338491296, + "grad_norm": 1.801928997039795, + "learning_rate": 0.00016582953765520686, + "loss": 0.1725, + "step": 32860 + }, + { + "epoch": 0.5127285206214514, + "grad_norm": 0.6406471729278564, + "learning_rate": 0.00016581913853705208, + "loss": 0.164, + "step": 32870 + }, + { + "epoch": 0.512884507393773, + "grad_norm": 0.9627325534820557, + "learning_rate": 0.0001658087394188973, + "loss": 0.1728, + "step": 32880 + }, + { + "epoch": 0.5130404941660948, + "grad_norm": 1.5446962118148804, + "learning_rate": 0.00016579834030074251, + "loss": 0.2345, + "step": 32890 + }, + { + "epoch": 0.5131964809384164, + "grad_norm": 0.28092706203460693, + "learning_rate": 0.0001657879411825877, + "loss": 0.2124, + "step": 32900 + }, + { + "epoch": 0.5133524677107382, + "grad_norm": 1.7751421928405762, + "learning_rate": 0.00016577754206443295, + "loss": 0.2367, + "step": 32910 + }, + { + "epoch": 0.5135084544830598, + "grad_norm": 1.6964634656906128, + "learning_rate": 0.00016576714294627815, + "loss": 0.2361, + "step": 32920 + }, + { + "epoch": 0.5136644412553816, + "grad_norm": 0.06759823113679886, + "learning_rate": 0.0001657567438281234, + "loss": 0.2329, + "step": 32930 + }, + { + "epoch": 0.5138204280277032, + "grad_norm": 2.2519280910491943, + "learning_rate": 0.00016574634470996859, + "loss": 0.1079, + "step": 32940 + }, + { + "epoch": 0.513976414800025, + "grad_norm": 3.167576551437378, + "learning_rate": 0.00016573594559181383, + "loss": 0.1764, + "step": 32950 + }, + { + "epoch": 0.5141324015723466, + "grad_norm": 1.6734397411346436, + "learning_rate": 0.00016572554647365903, + "loss": 0.2878, + "step": 32960 + }, + { + "epoch": 0.5142883883446684, + "grad_norm": 0.35627374053001404, + "learning_rate": 0.00016571514735550427, + "loss": 0.2337, + "step": 32970 + }, + { + "epoch": 0.51444437511699, + "grad_norm": 1.6321039199829102, + "learning_rate": 0.00016570474823734946, + "loss": 0.2551, + "step": 32980 + }, + { + "epoch": 0.5146003618893118, + "grad_norm": 1.352982997894287, + "learning_rate": 0.0001656943491191947, + "loss": 0.4467, + "step": 32990 + }, + { + "epoch": 0.5147563486616334, + "grad_norm": 0.3558129072189331, + "learning_rate": 0.0001656839500010399, + "loss": 0.1913, + "step": 33000 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.009561038848e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6bd09a1dad488d63af91f953c7c1dc46c90e6772 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10d0d336e6fcb44b9768dea63b9d035c77386f4434f3e865fd6b7c487f497471 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..d3bedb5663b8f17e5a501364f972ec774b6645bf --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5ab89d6e62c90213212ccf587316807748ba11ed53965b925f45f99fd5f312b +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..aad9364f5249c670402a54157d6831a0f9a6a454 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..f2d94476bf6c79aa2d0276970233ed210f076eab Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..b78e641052c789c7c41de6cc4230187759df9084 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/trainer_state.json @@ -0,0 +1,23484 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5225556872777188, + "eval_steps": 500, + "global_step": 33500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + }, + { + "epoch": 0.5071129968178698, + "grad_norm": 1.0075199604034424, + "learning_rate": 0.00016619350679062414, + "loss": 0.2569, + "step": 32510 + }, + { + "epoch": 0.5072689835901916, + "grad_norm": 1.3489066362380981, + "learning_rate": 0.0001661831076724694, + "loss": 0.2565, + "step": 32520 + }, + { + "epoch": 0.5074249703625132, + "grad_norm": 0.2534058094024658, + "learning_rate": 0.00016617270855431458, + "loss": 0.2292, + "step": 32530 + }, + { + "epoch": 0.507580957134835, + "grad_norm": 0.31723281741142273, + "learning_rate": 0.00016616230943615983, + "loss": 0.1467, + "step": 32540 + }, + { + "epoch": 0.5077369439071566, + "grad_norm": 0.29659849405288696, + "learning_rate": 0.00016615191031800502, + "loss": 0.2768, + "step": 32550 + }, + { + "epoch": 0.5078929306794784, + "grad_norm": 1.5246435403823853, + "learning_rate": 0.00016614151119985027, + "loss": 0.216, + "step": 32560 + }, + { + "epoch": 0.5080489174518001, + "grad_norm": 0.6026854515075684, + "learning_rate": 0.00016613111208169546, + "loss": 0.2277, + "step": 32570 + }, + { + "epoch": 0.5082049042241218, + "grad_norm": 0.41994890570640564, + "learning_rate": 0.0001661207129635407, + "loss": 0.2433, + "step": 32580 + }, + { + "epoch": 0.5083608909964435, + "grad_norm": 2.084599733352661, + "learning_rate": 0.00016611031384538593, + "loss": 0.4081, + "step": 32590 + }, + { + "epoch": 0.5085168777687652, + "grad_norm": 0.2945229709148407, + "learning_rate": 0.00016609991472723115, + "loss": 0.1329, + "step": 32600 + }, + { + "epoch": 0.508672864541087, + "grad_norm": 1.4981423616409302, + "learning_rate": 0.00016608951560907636, + "loss": 0.2855, + "step": 32610 + }, + { + "epoch": 0.5088288513134086, + "grad_norm": 0.961737871170044, + "learning_rate": 0.00016607911649092158, + "loss": 0.1654, + "step": 32620 + }, + { + "epoch": 0.5089848380857304, + "grad_norm": 1.49734628200531, + "learning_rate": 0.0001660687173727668, + "loss": 0.2081, + "step": 32630 + }, + { + "epoch": 0.509140824858052, + "grad_norm": 2.538135051727295, + "learning_rate": 0.00016605831825461202, + "loss": 0.3023, + "step": 32640 + }, + { + "epoch": 0.5092968116303738, + "grad_norm": 0.2981272041797638, + "learning_rate": 0.00016604791913645724, + "loss": 0.214, + "step": 32650 + }, + { + "epoch": 0.5094527984026954, + "grad_norm": 1.8863650560379028, + "learning_rate": 0.00016603752001830246, + "loss": 0.3363, + "step": 32660 + }, + { + "epoch": 0.5096087851750172, + "grad_norm": 0.022720983251929283, + "learning_rate": 0.00016602712090014768, + "loss": 0.2463, + "step": 32670 + }, + { + "epoch": 0.5097647719473388, + "grad_norm": 1.154977798461914, + "learning_rate": 0.0001660167217819929, + "loss": 0.4484, + "step": 32680 + }, + { + "epoch": 0.5099207587196606, + "grad_norm": 0.9470499157905579, + "learning_rate": 0.00016600632266383812, + "loss": 0.1196, + "step": 32690 + }, + { + "epoch": 0.5100767454919822, + "grad_norm": 0.19955401122570038, + "learning_rate": 0.00016599592354568334, + "loss": 0.1762, + "step": 32700 + }, + { + "epoch": 0.510232732264304, + "grad_norm": 0.4098282754421234, + "learning_rate": 0.00016598552442752856, + "loss": 0.2739, + "step": 32710 + }, + { + "epoch": 0.5103887190366257, + "grad_norm": 0.21697258949279785, + "learning_rate": 0.00016597512530937378, + "loss": 0.2555, + "step": 32720 + }, + { + "epoch": 0.5105447058089474, + "grad_norm": 4.894479751586914, + "learning_rate": 0.000165964726191219, + "loss": 0.3137, + "step": 32730 + }, + { + "epoch": 0.5107006925812692, + "grad_norm": 0.8396711349487305, + "learning_rate": 0.00016595432707306422, + "loss": 0.3105, + "step": 32740 + }, + { + "epoch": 0.5108566793535908, + "grad_norm": 1.5494674444198608, + "learning_rate": 0.00016594392795490944, + "loss": 0.2271, + "step": 32750 + }, + { + "epoch": 0.5110126661259126, + "grad_norm": 1.3409888744354248, + "learning_rate": 0.00016593352883675466, + "loss": 0.1796, + "step": 32760 + }, + { + "epoch": 0.5111686528982342, + "grad_norm": 1.7590911388397217, + "learning_rate": 0.00016592312971859988, + "loss": 0.2185, + "step": 32770 + }, + { + "epoch": 0.511324639670556, + "grad_norm": 0.7350002527236938, + "learning_rate": 0.0001659127306004451, + "loss": 0.1911, + "step": 32780 + }, + { + "epoch": 0.5114806264428776, + "grad_norm": 0.7176411747932434, + "learning_rate": 0.00016590233148229032, + "loss": 0.6918, + "step": 32790 + }, + { + "epoch": 0.5116366132151994, + "grad_norm": 2.5005321502685547, + "learning_rate": 0.00016589193236413554, + "loss": 0.2329, + "step": 32800 + }, + { + "epoch": 0.511792599987521, + "grad_norm": 1.3916250467300415, + "learning_rate": 0.00016588153324598076, + "loss": 0.2102, + "step": 32810 + }, + { + "epoch": 0.5119485867598428, + "grad_norm": 0.721671998500824, + "learning_rate": 0.00016587113412782598, + "loss": 0.2965, + "step": 32820 + }, + { + "epoch": 0.5121045735321644, + "grad_norm": 0.6120114326477051, + "learning_rate": 0.0001658607350096712, + "loss": 0.2661, + "step": 32830 + }, + { + "epoch": 0.5122605603044862, + "grad_norm": 0.29123347997665405, + "learning_rate": 0.00016585033589151642, + "loss": 0.2455, + "step": 32840 + }, + { + "epoch": 0.5124165470768078, + "grad_norm": 2.7300868034362793, + "learning_rate": 0.00016583993677336164, + "loss": 0.2049, + "step": 32850 + }, + { + "epoch": 0.5125725338491296, + "grad_norm": 1.801928997039795, + "learning_rate": 0.00016582953765520686, + "loss": 0.1725, + "step": 32860 + }, + { + "epoch": 0.5127285206214514, + "grad_norm": 0.6406471729278564, + "learning_rate": 0.00016581913853705208, + "loss": 0.164, + "step": 32870 + }, + { + "epoch": 0.512884507393773, + "grad_norm": 0.9627325534820557, + "learning_rate": 0.0001658087394188973, + "loss": 0.1728, + "step": 32880 + }, + { + "epoch": 0.5130404941660948, + "grad_norm": 1.5446962118148804, + "learning_rate": 0.00016579834030074251, + "loss": 0.2345, + "step": 32890 + }, + { + "epoch": 0.5131964809384164, + "grad_norm": 0.28092706203460693, + "learning_rate": 0.0001657879411825877, + "loss": 0.2124, + "step": 32900 + }, + { + "epoch": 0.5133524677107382, + "grad_norm": 1.7751421928405762, + "learning_rate": 0.00016577754206443295, + "loss": 0.2367, + "step": 32910 + }, + { + "epoch": 0.5135084544830598, + "grad_norm": 1.6964634656906128, + "learning_rate": 0.00016576714294627815, + "loss": 0.2361, + "step": 32920 + }, + { + "epoch": 0.5136644412553816, + "grad_norm": 0.06759823113679886, + "learning_rate": 0.0001657567438281234, + "loss": 0.2329, + "step": 32930 + }, + { + "epoch": 0.5138204280277032, + "grad_norm": 2.2519280910491943, + "learning_rate": 0.00016574634470996859, + "loss": 0.1079, + "step": 32940 + }, + { + "epoch": 0.513976414800025, + "grad_norm": 3.167576551437378, + "learning_rate": 0.00016573594559181383, + "loss": 0.1764, + "step": 32950 + }, + { + "epoch": 0.5141324015723466, + "grad_norm": 1.6734397411346436, + "learning_rate": 0.00016572554647365903, + "loss": 0.2878, + "step": 32960 + }, + { + "epoch": 0.5142883883446684, + "grad_norm": 0.35627374053001404, + "learning_rate": 0.00016571514735550427, + "loss": 0.2337, + "step": 32970 + }, + { + "epoch": 0.51444437511699, + "grad_norm": 1.6321039199829102, + "learning_rate": 0.00016570474823734946, + "loss": 0.2551, + "step": 32980 + }, + { + "epoch": 0.5146003618893118, + "grad_norm": 1.352982997894287, + "learning_rate": 0.0001656943491191947, + "loss": 0.4467, + "step": 32990 + }, + { + "epoch": 0.5147563486616334, + "grad_norm": 0.3558129072189331, + "learning_rate": 0.0001656839500010399, + "loss": 0.1913, + "step": 33000 + }, + { + "epoch": 0.5149123354339552, + "grad_norm": 0.041972286999225616, + "learning_rate": 0.00016567355088288515, + "loss": 0.1704, + "step": 33010 + }, + { + "epoch": 0.515068322206277, + "grad_norm": 1.4168939590454102, + "learning_rate": 0.00016566315176473034, + "loss": 0.4866, + "step": 33020 + }, + { + "epoch": 0.5152243089785986, + "grad_norm": 3.0042929649353027, + "learning_rate": 0.0001656527526465756, + "loss": 0.2473, + "step": 33030 + }, + { + "epoch": 0.5153802957509204, + "grad_norm": 1.027274489402771, + "learning_rate": 0.00016564235352842078, + "loss": 0.436, + "step": 33040 + }, + { + "epoch": 0.515536282523242, + "grad_norm": 1.2330195903778076, + "learning_rate": 0.00016563195441026603, + "loss": 0.4615, + "step": 33050 + }, + { + "epoch": 0.5156922692955638, + "grad_norm": 2.7145066261291504, + "learning_rate": 0.00016562155529211122, + "loss": 0.278, + "step": 33060 + }, + { + "epoch": 0.5158482560678854, + "grad_norm": 1.1273893117904663, + "learning_rate": 0.00016561115617395647, + "loss": 0.3767, + "step": 33070 + }, + { + "epoch": 0.5160042428402072, + "grad_norm": 2.466265916824341, + "learning_rate": 0.00016560075705580166, + "loss": 0.1879, + "step": 33080 + }, + { + "epoch": 0.5161602296125288, + "grad_norm": 0.3759515583515167, + "learning_rate": 0.0001655903579376469, + "loss": 0.4004, + "step": 33090 + }, + { + "epoch": 0.5163162163848506, + "grad_norm": 1.8821336030960083, + "learning_rate": 0.0001655799588194921, + "loss": 0.1622, + "step": 33100 + }, + { + "epoch": 0.5164722031571722, + "grad_norm": 0.8162096738815308, + "learning_rate": 0.00016556955970133735, + "loss": 0.1341, + "step": 33110 + }, + { + "epoch": 0.516628189929494, + "grad_norm": 0.008498801849782467, + "learning_rate": 0.00016555916058318254, + "loss": 0.2546, + "step": 33120 + }, + { + "epoch": 0.5167841767018156, + "grad_norm": 0.17536675930023193, + "learning_rate": 0.00016554876146502779, + "loss": 0.2934, + "step": 33130 + }, + { + "epoch": 0.5169401634741374, + "grad_norm": 3.539898633956909, + "learning_rate": 0.00016553836234687298, + "loss": 0.4959, + "step": 33140 + }, + { + "epoch": 0.517096150246459, + "grad_norm": 2.190887212753296, + "learning_rate": 0.00016552796322871823, + "loss": 0.2914, + "step": 33150 + }, + { + "epoch": 0.5172521370187808, + "grad_norm": 1.6388943195343018, + "learning_rate": 0.00016551756411056342, + "loss": 0.1317, + "step": 33160 + }, + { + "epoch": 0.5174081237911026, + "grad_norm": 0.33063387870788574, + "learning_rate": 0.00016550716499240866, + "loss": 0.4258, + "step": 33170 + }, + { + "epoch": 0.5175641105634242, + "grad_norm": 0.5168288946151733, + "learning_rate": 0.00016549676587425386, + "loss": 0.2548, + "step": 33180 + }, + { + "epoch": 0.517720097335746, + "grad_norm": 3.9570281505584717, + "learning_rate": 0.0001654863667560991, + "loss": 0.3179, + "step": 33190 + }, + { + "epoch": 0.5178760841080676, + "grad_norm": 1.127255916595459, + "learning_rate": 0.0001654759676379443, + "loss": 0.233, + "step": 33200 + }, + { + "epoch": 0.5180320708803894, + "grad_norm": 0.36299410462379456, + "learning_rate": 0.00016546556851978954, + "loss": 0.4614, + "step": 33210 + }, + { + "epoch": 0.518188057652711, + "grad_norm": 2.914668321609497, + "learning_rate": 0.00016545516940163474, + "loss": 0.1603, + "step": 33220 + }, + { + "epoch": 0.5183440444250328, + "grad_norm": 6.991094589233398, + "learning_rate": 0.00016544477028347998, + "loss": 0.2259, + "step": 33230 + }, + { + "epoch": 0.5185000311973544, + "grad_norm": 2.0594241619110107, + "learning_rate": 0.00016543437116532517, + "loss": 0.1885, + "step": 33240 + }, + { + "epoch": 0.5186560179696762, + "grad_norm": 0.5594693422317505, + "learning_rate": 0.00016542397204717042, + "loss": 0.4333, + "step": 33250 + }, + { + "epoch": 0.5188120047419978, + "grad_norm": 4.314854621887207, + "learning_rate": 0.00016541357292901561, + "loss": 0.2103, + "step": 33260 + }, + { + "epoch": 0.5189679915143196, + "grad_norm": 0.5609000325202942, + "learning_rate": 0.00016540317381086086, + "loss": 0.2498, + "step": 33270 + }, + { + "epoch": 0.5191239782866413, + "grad_norm": 3.680088758468628, + "learning_rate": 0.00016539277469270605, + "loss": 0.3465, + "step": 33280 + }, + { + "epoch": 0.519279965058963, + "grad_norm": 1.4993672370910645, + "learning_rate": 0.0001653823755745513, + "loss": 0.3272, + "step": 33290 + }, + { + "epoch": 0.5194359518312847, + "grad_norm": 0.18005084991455078, + "learning_rate": 0.0001653719764563965, + "loss": 0.1706, + "step": 33300 + }, + { + "epoch": 0.5195919386036064, + "grad_norm": 0.5105524659156799, + "learning_rate": 0.00016536157733824174, + "loss": 0.0937, + "step": 33310 + }, + { + "epoch": 0.5197479253759281, + "grad_norm": 1.2806589603424072, + "learning_rate": 0.00016535117822008693, + "loss": 0.1578, + "step": 33320 + }, + { + "epoch": 0.5199039121482498, + "grad_norm": 0.22015990316867828, + "learning_rate": 0.00016534077910193218, + "loss": 0.2081, + "step": 33330 + }, + { + "epoch": 0.5200598989205716, + "grad_norm": 0.34638991951942444, + "learning_rate": 0.00016533037998377737, + "loss": 0.2389, + "step": 33340 + }, + { + "epoch": 0.5202158856928932, + "grad_norm": 0.8709302544593811, + "learning_rate": 0.0001653199808656226, + "loss": 0.2112, + "step": 33350 + }, + { + "epoch": 0.520371872465215, + "grad_norm": 1.2309521436691284, + "learning_rate": 0.0001653095817474678, + "loss": 0.3448, + "step": 33360 + }, + { + "epoch": 0.5205278592375366, + "grad_norm": 1.3155115842819214, + "learning_rate": 0.00016529918262931303, + "loss": 0.1572, + "step": 33370 + }, + { + "epoch": 0.5206838460098584, + "grad_norm": 0.358751118183136, + "learning_rate": 0.00016528878351115825, + "loss": 0.3011, + "step": 33380 + }, + { + "epoch": 0.52083983278218, + "grad_norm": 3.800628662109375, + "learning_rate": 0.00016527838439300347, + "loss": 0.458, + "step": 33390 + }, + { + "epoch": 0.5209958195545018, + "grad_norm": 1.1253124475479126, + "learning_rate": 0.0001652679852748487, + "loss": 0.1238, + "step": 33400 + }, + { + "epoch": 0.5211518063268235, + "grad_norm": 0.46979108452796936, + "learning_rate": 0.0001652575861566939, + "loss": 0.0745, + "step": 33410 + }, + { + "epoch": 0.5213077930991452, + "grad_norm": 7.921492576599121, + "learning_rate": 0.00016524718703853913, + "loss": 0.4004, + "step": 33420 + }, + { + "epoch": 0.5214637798714669, + "grad_norm": 1.643127679824829, + "learning_rate": 0.00016523678792038435, + "loss": 0.2858, + "step": 33430 + }, + { + "epoch": 0.5216197666437886, + "grad_norm": 7.096631050109863, + "learning_rate": 0.0001652263888022296, + "loss": 0.4735, + "step": 33440 + }, + { + "epoch": 0.5217757534161103, + "grad_norm": 0.7280219197273254, + "learning_rate": 0.0001652159896840748, + "loss": 0.1841, + "step": 33450 + }, + { + "epoch": 0.521931740188432, + "grad_norm": 0.9415522813796997, + "learning_rate": 0.00016520559056592003, + "loss": 0.2169, + "step": 33460 + }, + { + "epoch": 0.5220877269607537, + "grad_norm": 2.264862060546875, + "learning_rate": 0.00016519519144776523, + "loss": 0.1902, + "step": 33470 + }, + { + "epoch": 0.5222437137330754, + "grad_norm": 2.4566822052001953, + "learning_rate": 0.00016518479232961047, + "loss": 0.3329, + "step": 33480 + }, + { + "epoch": 0.5223997005053972, + "grad_norm": 0.7330378890037537, + "learning_rate": 0.00016517439321145567, + "loss": 0.2341, + "step": 33490 + }, + { + "epoch": 0.5225556872777188, + "grad_norm": 1.6409542560577393, + "learning_rate": 0.0001651639940933009, + "loss": 0.3077, + "step": 33500 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.040008933376e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-33500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..52666babf174d0bbc92f10ae1044f72d380ef661 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a1322b8fb614a817d239d37d2023229effb10b511411d82d65b97aa1044e3e80 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..90550df3e8def3f1622f08018943530a4e8b8faa --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3963109cbd2ee4cdd6f1b551c1b4947ef9495ff8197041e26089b687949e7a8 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..c2f17766f88e644ed45ebacff8b740c8a7de22ea Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..83253fe4e8fb96ad021280cc7cc5a87c1a4b4742 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..da69bdd9c71ef4018efe70ab2e7fe5983d6e4a06 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/trainer_state.json @@ -0,0 +1,23834 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5303550258938042, + "eval_steps": 500, + "global_step": 34000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + }, + { + "epoch": 0.5071129968178698, + "grad_norm": 1.0075199604034424, + "learning_rate": 0.00016619350679062414, + "loss": 0.2569, + "step": 32510 + }, + { + "epoch": 0.5072689835901916, + "grad_norm": 1.3489066362380981, + "learning_rate": 0.0001661831076724694, + "loss": 0.2565, + "step": 32520 + }, + { + "epoch": 0.5074249703625132, + "grad_norm": 0.2534058094024658, + "learning_rate": 0.00016617270855431458, + "loss": 0.2292, + "step": 32530 + }, + { + "epoch": 0.507580957134835, + "grad_norm": 0.31723281741142273, + "learning_rate": 0.00016616230943615983, + "loss": 0.1467, + "step": 32540 + }, + { + "epoch": 0.5077369439071566, + "grad_norm": 0.29659849405288696, + "learning_rate": 0.00016615191031800502, + "loss": 0.2768, + "step": 32550 + }, + { + "epoch": 0.5078929306794784, + "grad_norm": 1.5246435403823853, + "learning_rate": 0.00016614151119985027, + "loss": 0.216, + "step": 32560 + }, + { + "epoch": 0.5080489174518001, + "grad_norm": 0.6026854515075684, + "learning_rate": 0.00016613111208169546, + "loss": 0.2277, + "step": 32570 + }, + { + "epoch": 0.5082049042241218, + "grad_norm": 0.41994890570640564, + "learning_rate": 0.0001661207129635407, + "loss": 0.2433, + "step": 32580 + }, + { + "epoch": 0.5083608909964435, + "grad_norm": 2.084599733352661, + "learning_rate": 0.00016611031384538593, + "loss": 0.4081, + "step": 32590 + }, + { + "epoch": 0.5085168777687652, + "grad_norm": 0.2945229709148407, + "learning_rate": 0.00016609991472723115, + "loss": 0.1329, + "step": 32600 + }, + { + "epoch": 0.508672864541087, + "grad_norm": 1.4981423616409302, + "learning_rate": 0.00016608951560907636, + "loss": 0.2855, + "step": 32610 + }, + { + "epoch": 0.5088288513134086, + "grad_norm": 0.961737871170044, + "learning_rate": 0.00016607911649092158, + "loss": 0.1654, + "step": 32620 + }, + { + "epoch": 0.5089848380857304, + "grad_norm": 1.49734628200531, + "learning_rate": 0.0001660687173727668, + "loss": 0.2081, + "step": 32630 + }, + { + "epoch": 0.509140824858052, + "grad_norm": 2.538135051727295, + "learning_rate": 0.00016605831825461202, + "loss": 0.3023, + "step": 32640 + }, + { + "epoch": 0.5092968116303738, + "grad_norm": 0.2981272041797638, + "learning_rate": 0.00016604791913645724, + "loss": 0.214, + "step": 32650 + }, + { + "epoch": 0.5094527984026954, + "grad_norm": 1.8863650560379028, + "learning_rate": 0.00016603752001830246, + "loss": 0.3363, + "step": 32660 + }, + { + "epoch": 0.5096087851750172, + "grad_norm": 0.022720983251929283, + "learning_rate": 0.00016602712090014768, + "loss": 0.2463, + "step": 32670 + }, + { + "epoch": 0.5097647719473388, + "grad_norm": 1.154977798461914, + "learning_rate": 0.0001660167217819929, + "loss": 0.4484, + "step": 32680 + }, + { + "epoch": 0.5099207587196606, + "grad_norm": 0.9470499157905579, + "learning_rate": 0.00016600632266383812, + "loss": 0.1196, + "step": 32690 + }, + { + "epoch": 0.5100767454919822, + "grad_norm": 0.19955401122570038, + "learning_rate": 0.00016599592354568334, + "loss": 0.1762, + "step": 32700 + }, + { + "epoch": 0.510232732264304, + "grad_norm": 0.4098282754421234, + "learning_rate": 0.00016598552442752856, + "loss": 0.2739, + "step": 32710 + }, + { + "epoch": 0.5103887190366257, + "grad_norm": 0.21697258949279785, + "learning_rate": 0.00016597512530937378, + "loss": 0.2555, + "step": 32720 + }, + { + "epoch": 0.5105447058089474, + "grad_norm": 4.894479751586914, + "learning_rate": 0.000165964726191219, + "loss": 0.3137, + "step": 32730 + }, + { + "epoch": 0.5107006925812692, + "grad_norm": 0.8396711349487305, + "learning_rate": 0.00016595432707306422, + "loss": 0.3105, + "step": 32740 + }, + { + "epoch": 0.5108566793535908, + "grad_norm": 1.5494674444198608, + "learning_rate": 0.00016594392795490944, + "loss": 0.2271, + "step": 32750 + }, + { + "epoch": 0.5110126661259126, + "grad_norm": 1.3409888744354248, + "learning_rate": 0.00016593352883675466, + "loss": 0.1796, + "step": 32760 + }, + { + "epoch": 0.5111686528982342, + "grad_norm": 1.7590911388397217, + "learning_rate": 0.00016592312971859988, + "loss": 0.2185, + "step": 32770 + }, + { + "epoch": 0.511324639670556, + "grad_norm": 0.7350002527236938, + "learning_rate": 0.0001659127306004451, + "loss": 0.1911, + "step": 32780 + }, + { + "epoch": 0.5114806264428776, + "grad_norm": 0.7176411747932434, + "learning_rate": 0.00016590233148229032, + "loss": 0.6918, + "step": 32790 + }, + { + "epoch": 0.5116366132151994, + "grad_norm": 2.5005321502685547, + "learning_rate": 0.00016589193236413554, + "loss": 0.2329, + "step": 32800 + }, + { + "epoch": 0.511792599987521, + "grad_norm": 1.3916250467300415, + "learning_rate": 0.00016588153324598076, + "loss": 0.2102, + "step": 32810 + }, + { + "epoch": 0.5119485867598428, + "grad_norm": 0.721671998500824, + "learning_rate": 0.00016587113412782598, + "loss": 0.2965, + "step": 32820 + }, + { + "epoch": 0.5121045735321644, + "grad_norm": 0.6120114326477051, + "learning_rate": 0.0001658607350096712, + "loss": 0.2661, + "step": 32830 + }, + { + "epoch": 0.5122605603044862, + "grad_norm": 0.29123347997665405, + "learning_rate": 0.00016585033589151642, + "loss": 0.2455, + "step": 32840 + }, + { + "epoch": 0.5124165470768078, + "grad_norm": 2.7300868034362793, + "learning_rate": 0.00016583993677336164, + "loss": 0.2049, + "step": 32850 + }, + { + "epoch": 0.5125725338491296, + "grad_norm": 1.801928997039795, + "learning_rate": 0.00016582953765520686, + "loss": 0.1725, + "step": 32860 + }, + { + "epoch": 0.5127285206214514, + "grad_norm": 0.6406471729278564, + "learning_rate": 0.00016581913853705208, + "loss": 0.164, + "step": 32870 + }, + { + "epoch": 0.512884507393773, + "grad_norm": 0.9627325534820557, + "learning_rate": 0.0001658087394188973, + "loss": 0.1728, + "step": 32880 + }, + { + "epoch": 0.5130404941660948, + "grad_norm": 1.5446962118148804, + "learning_rate": 0.00016579834030074251, + "loss": 0.2345, + "step": 32890 + }, + { + "epoch": 0.5131964809384164, + "grad_norm": 0.28092706203460693, + "learning_rate": 0.0001657879411825877, + "loss": 0.2124, + "step": 32900 + }, + { + "epoch": 0.5133524677107382, + "grad_norm": 1.7751421928405762, + "learning_rate": 0.00016577754206443295, + "loss": 0.2367, + "step": 32910 + }, + { + "epoch": 0.5135084544830598, + "grad_norm": 1.6964634656906128, + "learning_rate": 0.00016576714294627815, + "loss": 0.2361, + "step": 32920 + }, + { + "epoch": 0.5136644412553816, + "grad_norm": 0.06759823113679886, + "learning_rate": 0.0001657567438281234, + "loss": 0.2329, + "step": 32930 + }, + { + "epoch": 0.5138204280277032, + "grad_norm": 2.2519280910491943, + "learning_rate": 0.00016574634470996859, + "loss": 0.1079, + "step": 32940 + }, + { + "epoch": 0.513976414800025, + "grad_norm": 3.167576551437378, + "learning_rate": 0.00016573594559181383, + "loss": 0.1764, + "step": 32950 + }, + { + "epoch": 0.5141324015723466, + "grad_norm": 1.6734397411346436, + "learning_rate": 0.00016572554647365903, + "loss": 0.2878, + "step": 32960 + }, + { + "epoch": 0.5142883883446684, + "grad_norm": 0.35627374053001404, + "learning_rate": 0.00016571514735550427, + "loss": 0.2337, + "step": 32970 + }, + { + "epoch": 0.51444437511699, + "grad_norm": 1.6321039199829102, + "learning_rate": 0.00016570474823734946, + "loss": 0.2551, + "step": 32980 + }, + { + "epoch": 0.5146003618893118, + "grad_norm": 1.352982997894287, + "learning_rate": 0.0001656943491191947, + "loss": 0.4467, + "step": 32990 + }, + { + "epoch": 0.5147563486616334, + "grad_norm": 0.3558129072189331, + "learning_rate": 0.0001656839500010399, + "loss": 0.1913, + "step": 33000 + }, + { + "epoch": 0.5149123354339552, + "grad_norm": 0.041972286999225616, + "learning_rate": 0.00016567355088288515, + "loss": 0.1704, + "step": 33010 + }, + { + "epoch": 0.515068322206277, + "grad_norm": 1.4168939590454102, + "learning_rate": 0.00016566315176473034, + "loss": 0.4866, + "step": 33020 + }, + { + "epoch": 0.5152243089785986, + "grad_norm": 3.0042929649353027, + "learning_rate": 0.0001656527526465756, + "loss": 0.2473, + "step": 33030 + }, + { + "epoch": 0.5153802957509204, + "grad_norm": 1.027274489402771, + "learning_rate": 0.00016564235352842078, + "loss": 0.436, + "step": 33040 + }, + { + "epoch": 0.515536282523242, + "grad_norm": 1.2330195903778076, + "learning_rate": 0.00016563195441026603, + "loss": 0.4615, + "step": 33050 + }, + { + "epoch": 0.5156922692955638, + "grad_norm": 2.7145066261291504, + "learning_rate": 0.00016562155529211122, + "loss": 0.278, + "step": 33060 + }, + { + "epoch": 0.5158482560678854, + "grad_norm": 1.1273893117904663, + "learning_rate": 0.00016561115617395647, + "loss": 0.3767, + "step": 33070 + }, + { + "epoch": 0.5160042428402072, + "grad_norm": 2.466265916824341, + "learning_rate": 0.00016560075705580166, + "loss": 0.1879, + "step": 33080 + }, + { + "epoch": 0.5161602296125288, + "grad_norm": 0.3759515583515167, + "learning_rate": 0.0001655903579376469, + "loss": 0.4004, + "step": 33090 + }, + { + "epoch": 0.5163162163848506, + "grad_norm": 1.8821336030960083, + "learning_rate": 0.0001655799588194921, + "loss": 0.1622, + "step": 33100 + }, + { + "epoch": 0.5164722031571722, + "grad_norm": 0.8162096738815308, + "learning_rate": 0.00016556955970133735, + "loss": 0.1341, + "step": 33110 + }, + { + "epoch": 0.516628189929494, + "grad_norm": 0.008498801849782467, + "learning_rate": 0.00016555916058318254, + "loss": 0.2546, + "step": 33120 + }, + { + "epoch": 0.5167841767018156, + "grad_norm": 0.17536675930023193, + "learning_rate": 0.00016554876146502779, + "loss": 0.2934, + "step": 33130 + }, + { + "epoch": 0.5169401634741374, + "grad_norm": 3.539898633956909, + "learning_rate": 0.00016553836234687298, + "loss": 0.4959, + "step": 33140 + }, + { + "epoch": 0.517096150246459, + "grad_norm": 2.190887212753296, + "learning_rate": 0.00016552796322871823, + "loss": 0.2914, + "step": 33150 + }, + { + "epoch": 0.5172521370187808, + "grad_norm": 1.6388943195343018, + "learning_rate": 0.00016551756411056342, + "loss": 0.1317, + "step": 33160 + }, + { + "epoch": 0.5174081237911026, + "grad_norm": 0.33063387870788574, + "learning_rate": 0.00016550716499240866, + "loss": 0.4258, + "step": 33170 + }, + { + "epoch": 0.5175641105634242, + "grad_norm": 0.5168288946151733, + "learning_rate": 0.00016549676587425386, + "loss": 0.2548, + "step": 33180 + }, + { + "epoch": 0.517720097335746, + "grad_norm": 3.9570281505584717, + "learning_rate": 0.0001654863667560991, + "loss": 0.3179, + "step": 33190 + }, + { + "epoch": 0.5178760841080676, + "grad_norm": 1.127255916595459, + "learning_rate": 0.0001654759676379443, + "loss": 0.233, + "step": 33200 + }, + { + "epoch": 0.5180320708803894, + "grad_norm": 0.36299410462379456, + "learning_rate": 0.00016546556851978954, + "loss": 0.4614, + "step": 33210 + }, + { + "epoch": 0.518188057652711, + "grad_norm": 2.914668321609497, + "learning_rate": 0.00016545516940163474, + "loss": 0.1603, + "step": 33220 + }, + { + "epoch": 0.5183440444250328, + "grad_norm": 6.991094589233398, + "learning_rate": 0.00016544477028347998, + "loss": 0.2259, + "step": 33230 + }, + { + "epoch": 0.5185000311973544, + "grad_norm": 2.0594241619110107, + "learning_rate": 0.00016543437116532517, + "loss": 0.1885, + "step": 33240 + }, + { + "epoch": 0.5186560179696762, + "grad_norm": 0.5594693422317505, + "learning_rate": 0.00016542397204717042, + "loss": 0.4333, + "step": 33250 + }, + { + "epoch": 0.5188120047419978, + "grad_norm": 4.314854621887207, + "learning_rate": 0.00016541357292901561, + "loss": 0.2103, + "step": 33260 + }, + { + "epoch": 0.5189679915143196, + "grad_norm": 0.5609000325202942, + "learning_rate": 0.00016540317381086086, + "loss": 0.2498, + "step": 33270 + }, + { + "epoch": 0.5191239782866413, + "grad_norm": 3.680088758468628, + "learning_rate": 0.00016539277469270605, + "loss": 0.3465, + "step": 33280 + }, + { + "epoch": 0.519279965058963, + "grad_norm": 1.4993672370910645, + "learning_rate": 0.0001653823755745513, + "loss": 0.3272, + "step": 33290 + }, + { + "epoch": 0.5194359518312847, + "grad_norm": 0.18005084991455078, + "learning_rate": 0.0001653719764563965, + "loss": 0.1706, + "step": 33300 + }, + { + "epoch": 0.5195919386036064, + "grad_norm": 0.5105524659156799, + "learning_rate": 0.00016536157733824174, + "loss": 0.0937, + "step": 33310 + }, + { + "epoch": 0.5197479253759281, + "grad_norm": 1.2806589603424072, + "learning_rate": 0.00016535117822008693, + "loss": 0.1578, + "step": 33320 + }, + { + "epoch": 0.5199039121482498, + "grad_norm": 0.22015990316867828, + "learning_rate": 0.00016534077910193218, + "loss": 0.2081, + "step": 33330 + }, + { + "epoch": 0.5200598989205716, + "grad_norm": 0.34638991951942444, + "learning_rate": 0.00016533037998377737, + "loss": 0.2389, + "step": 33340 + }, + { + "epoch": 0.5202158856928932, + "grad_norm": 0.8709302544593811, + "learning_rate": 0.0001653199808656226, + "loss": 0.2112, + "step": 33350 + }, + { + "epoch": 0.520371872465215, + "grad_norm": 1.2309521436691284, + "learning_rate": 0.0001653095817474678, + "loss": 0.3448, + "step": 33360 + }, + { + "epoch": 0.5205278592375366, + "grad_norm": 1.3155115842819214, + "learning_rate": 0.00016529918262931303, + "loss": 0.1572, + "step": 33370 + }, + { + "epoch": 0.5206838460098584, + "grad_norm": 0.358751118183136, + "learning_rate": 0.00016528878351115825, + "loss": 0.3011, + "step": 33380 + }, + { + "epoch": 0.52083983278218, + "grad_norm": 3.800628662109375, + "learning_rate": 0.00016527838439300347, + "loss": 0.458, + "step": 33390 + }, + { + "epoch": 0.5209958195545018, + "grad_norm": 1.1253124475479126, + "learning_rate": 0.0001652679852748487, + "loss": 0.1238, + "step": 33400 + }, + { + "epoch": 0.5211518063268235, + "grad_norm": 0.46979108452796936, + "learning_rate": 0.0001652575861566939, + "loss": 0.0745, + "step": 33410 + }, + { + "epoch": 0.5213077930991452, + "grad_norm": 7.921492576599121, + "learning_rate": 0.00016524718703853913, + "loss": 0.4004, + "step": 33420 + }, + { + "epoch": 0.5214637798714669, + "grad_norm": 1.643127679824829, + "learning_rate": 0.00016523678792038435, + "loss": 0.2858, + "step": 33430 + }, + { + "epoch": 0.5216197666437886, + "grad_norm": 7.096631050109863, + "learning_rate": 0.0001652263888022296, + "loss": 0.4735, + "step": 33440 + }, + { + "epoch": 0.5217757534161103, + "grad_norm": 0.7280219197273254, + "learning_rate": 0.0001652159896840748, + "loss": 0.1841, + "step": 33450 + }, + { + "epoch": 0.521931740188432, + "grad_norm": 0.9415522813796997, + "learning_rate": 0.00016520559056592003, + "loss": 0.2169, + "step": 33460 + }, + { + "epoch": 0.5220877269607537, + "grad_norm": 2.264862060546875, + "learning_rate": 0.00016519519144776523, + "loss": 0.1902, + "step": 33470 + }, + { + "epoch": 0.5222437137330754, + "grad_norm": 2.4566822052001953, + "learning_rate": 0.00016518479232961047, + "loss": 0.3329, + "step": 33480 + }, + { + "epoch": 0.5223997005053972, + "grad_norm": 0.7330378890037537, + "learning_rate": 0.00016517439321145567, + "loss": 0.2341, + "step": 33490 + }, + { + "epoch": 0.5225556872777188, + "grad_norm": 1.6409542560577393, + "learning_rate": 0.0001651639940933009, + "loss": 0.3077, + "step": 33500 + }, + { + "epoch": 0.5227116740500406, + "grad_norm": 1.0647432804107666, + "learning_rate": 0.0001651535949751461, + "loss": 0.1935, + "step": 33510 + }, + { + "epoch": 0.5228676608223622, + "grad_norm": 0.507420539855957, + "learning_rate": 0.00016514319585699135, + "loss": 0.1314, + "step": 33520 + }, + { + "epoch": 0.523023647594684, + "grad_norm": 1.0064164400100708, + "learning_rate": 0.00016513279673883654, + "loss": 0.344, + "step": 33530 + }, + { + "epoch": 0.5231796343670057, + "grad_norm": 1.1936209201812744, + "learning_rate": 0.0001651223976206818, + "loss": 0.1948, + "step": 33540 + }, + { + "epoch": 0.5233356211393274, + "grad_norm": 0.04992926865816116, + "learning_rate": 0.00016511199850252698, + "loss": 0.1363, + "step": 33550 + }, + { + "epoch": 0.5234916079116491, + "grad_norm": 1.8256275653839111, + "learning_rate": 0.00016510159938437223, + "loss": 0.443, + "step": 33560 + }, + { + "epoch": 0.5236475946839708, + "grad_norm": 4.914908409118652, + "learning_rate": 0.00016509120026621742, + "loss": 0.4654, + "step": 33570 + }, + { + "epoch": 0.5238035814562925, + "grad_norm": 1.1625486612319946, + "learning_rate": 0.00016508080114806267, + "loss": 0.1974, + "step": 33580 + }, + { + "epoch": 0.5239595682286142, + "grad_norm": 6.6192522048950195, + "learning_rate": 0.00016507040202990786, + "loss": 0.2131, + "step": 33590 + }, + { + "epoch": 0.5241155550009359, + "grad_norm": 0.7369065880775452, + "learning_rate": 0.0001650600029117531, + "loss": 0.2371, + "step": 33600 + }, + { + "epoch": 0.5242715417732576, + "grad_norm": 1.5238152742385864, + "learning_rate": 0.0001650496037935983, + "loss": 0.1557, + "step": 33610 + }, + { + "epoch": 0.5244275285455793, + "grad_norm": 1.0418007373809814, + "learning_rate": 0.00016503920467544355, + "loss": 0.1878, + "step": 33620 + }, + { + "epoch": 0.524583515317901, + "grad_norm": 0.790117084980011, + "learning_rate": 0.00016502880555728874, + "loss": 0.2195, + "step": 33630 + }, + { + "epoch": 0.5247395020902228, + "grad_norm": 1.6712257862091064, + "learning_rate": 0.000165018406439134, + "loss": 0.1602, + "step": 33640 + }, + { + "epoch": 0.5248954888625444, + "grad_norm": 0.19236230850219727, + "learning_rate": 0.00016500800732097918, + "loss": 0.2526, + "step": 33650 + }, + { + "epoch": 0.5250514756348662, + "grad_norm": 1.3519701957702637, + "learning_rate": 0.00016499760820282443, + "loss": 0.4686, + "step": 33660 + }, + { + "epoch": 0.5252074624071879, + "grad_norm": 1.694342851638794, + "learning_rate": 0.00016498720908466962, + "loss": 0.1859, + "step": 33670 + }, + { + "epoch": 0.5253634491795096, + "grad_norm": 5.225239276885986, + "learning_rate": 0.00016497680996651487, + "loss": 0.2944, + "step": 33680 + }, + { + "epoch": 0.5255194359518313, + "grad_norm": 2.0208842754364014, + "learning_rate": 0.00016496641084836006, + "loss": 0.2421, + "step": 33690 + }, + { + "epoch": 0.525675422724153, + "grad_norm": 0.7954996228218079, + "learning_rate": 0.0001649560117302053, + "loss": 0.2853, + "step": 33700 + }, + { + "epoch": 0.5258314094964747, + "grad_norm": 2.296086072921753, + "learning_rate": 0.0001649456126120505, + "loss": 0.1333, + "step": 33710 + }, + { + "epoch": 0.5259873962687964, + "grad_norm": 1.1779128313064575, + "learning_rate": 0.00016493521349389574, + "loss": 0.1066, + "step": 33720 + }, + { + "epoch": 0.5261433830411181, + "grad_norm": 0.1756065934896469, + "learning_rate": 0.00016492481437574094, + "loss": 0.1352, + "step": 33730 + }, + { + "epoch": 0.5262993698134398, + "grad_norm": 0.13100725412368774, + "learning_rate": 0.00016491441525758618, + "loss": 0.2399, + "step": 33740 + }, + { + "epoch": 0.5264553565857615, + "grad_norm": 5.532008171081543, + "learning_rate": 0.00016490401613943138, + "loss": 0.2896, + "step": 33750 + }, + { + "epoch": 0.5266113433580832, + "grad_norm": 1.319886565208435, + "learning_rate": 0.00016489361702127662, + "loss": 0.3275, + "step": 33760 + }, + { + "epoch": 0.5267673301304049, + "grad_norm": 1.5550974607467651, + "learning_rate": 0.00016488321790312182, + "loss": 0.2677, + "step": 33770 + }, + { + "epoch": 0.5269233169027266, + "grad_norm": 1.8936737775802612, + "learning_rate": 0.00016487281878496706, + "loss": 0.1955, + "step": 33780 + }, + { + "epoch": 0.5270793036750484, + "grad_norm": 0.3653401732444763, + "learning_rate": 0.00016486241966681226, + "loss": 0.0723, + "step": 33790 + }, + { + "epoch": 0.52723529044737, + "grad_norm": 2.861341714859009, + "learning_rate": 0.00016485202054865747, + "loss": 0.2412, + "step": 33800 + }, + { + "epoch": 0.5273912772196918, + "grad_norm": 1.5291428565979004, + "learning_rate": 0.0001648416214305027, + "loss": 0.0871, + "step": 33810 + }, + { + "epoch": 0.5275472639920135, + "grad_norm": 1.0372581481933594, + "learning_rate": 0.00016483122231234791, + "loss": 0.4705, + "step": 33820 + }, + { + "epoch": 0.5277032507643352, + "grad_norm": 1.1943141222000122, + "learning_rate": 0.00016482082319419313, + "loss": 0.2848, + "step": 33830 + }, + { + "epoch": 0.5278592375366569, + "grad_norm": 1.9008225202560425, + "learning_rate": 0.00016481042407603835, + "loss": 0.139, + "step": 33840 + }, + { + "epoch": 0.5280152243089786, + "grad_norm": 2.132089138031006, + "learning_rate": 0.00016480002495788357, + "loss": 0.2119, + "step": 33850 + }, + { + "epoch": 0.5281712110813003, + "grad_norm": 0.24524426460266113, + "learning_rate": 0.0001647896258397288, + "loss": 0.1643, + "step": 33860 + }, + { + "epoch": 0.528327197853622, + "grad_norm": 1.6469637155532837, + "learning_rate": 0.000164779226721574, + "loss": 0.363, + "step": 33870 + }, + { + "epoch": 0.5284831846259437, + "grad_norm": 0.8767328858375549, + "learning_rate": 0.00016476882760341923, + "loss": 0.2632, + "step": 33880 + }, + { + "epoch": 0.5286391713982654, + "grad_norm": 0.06347586214542389, + "learning_rate": 0.00016475842848526445, + "loss": 0.3204, + "step": 33890 + }, + { + "epoch": 0.5287951581705871, + "grad_norm": 0.09782540798187256, + "learning_rate": 0.00016474802936710967, + "loss": 0.2211, + "step": 33900 + }, + { + "epoch": 0.5289511449429088, + "grad_norm": 3.2998859882354736, + "learning_rate": 0.0001647376302489549, + "loss": 0.265, + "step": 33910 + }, + { + "epoch": 0.5291071317152305, + "grad_norm": 0.43594226241111755, + "learning_rate": 0.0001647272311308001, + "loss": 0.194, + "step": 33920 + }, + { + "epoch": 0.5292631184875523, + "grad_norm": 1.5166605710983276, + "learning_rate": 0.00016471683201264533, + "loss": 0.2675, + "step": 33930 + }, + { + "epoch": 0.529419105259874, + "grad_norm": 0.6056640148162842, + "learning_rate": 0.00016470643289449055, + "loss": 0.0576, + "step": 33940 + }, + { + "epoch": 0.5295750920321957, + "grad_norm": 0.25410348176956177, + "learning_rate": 0.00016469603377633577, + "loss": 0.1124, + "step": 33950 + }, + { + "epoch": 0.5297310788045174, + "grad_norm": 1.770642876625061, + "learning_rate": 0.000164685634658181, + "loss": 0.3295, + "step": 33960 + }, + { + "epoch": 0.5298870655768391, + "grad_norm": 0.0607205331325531, + "learning_rate": 0.0001646752355400262, + "loss": 0.2369, + "step": 33970 + }, + { + "epoch": 0.5300430523491608, + "grad_norm": 0.5557095408439636, + "learning_rate": 0.00016466483642187143, + "loss": 0.1681, + "step": 33980 + }, + { + "epoch": 0.5301990391214825, + "grad_norm": 0.5192957520484924, + "learning_rate": 0.00016465443730371665, + "loss": 0.2649, + "step": 33990 + }, + { + "epoch": 0.5303550258938042, + "grad_norm": 0.04804835096001625, + "learning_rate": 0.00016464403818556187, + "loss": 0.2639, + "step": 34000 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.070456827904e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e164a1a95f0b103537e95167c8744b71cbd2586d --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88dcb63f7c5ac0f5aa915422eb1ebcb864b79f535d0470fe98940d7bd9e4d754 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..42fa7d581935886b7ecf49ad592367f393d6743e --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d32c7167b7e31f58f3ecebf13043175347a9286b87693e8ce3e9fc61e28a465 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..f201b9a05cf3029055bb0b176ab8cc2105f1db89 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..fa0d7e25de9cf266166e692de96d76515d7d14d8 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..e87c010c3ec8c0e313ef83ac87d38556c264250e --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/trainer_state.json @@ -0,0 +1,24184 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5381543645098895, + "eval_steps": 500, + "global_step": 34500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + }, + { + "epoch": 0.5071129968178698, + "grad_norm": 1.0075199604034424, + "learning_rate": 0.00016619350679062414, + "loss": 0.2569, + "step": 32510 + }, + { + "epoch": 0.5072689835901916, + "grad_norm": 1.3489066362380981, + "learning_rate": 0.0001661831076724694, + "loss": 0.2565, + "step": 32520 + }, + { + "epoch": 0.5074249703625132, + "grad_norm": 0.2534058094024658, + "learning_rate": 0.00016617270855431458, + "loss": 0.2292, + "step": 32530 + }, + { + "epoch": 0.507580957134835, + "grad_norm": 0.31723281741142273, + "learning_rate": 0.00016616230943615983, + "loss": 0.1467, + "step": 32540 + }, + { + "epoch": 0.5077369439071566, + "grad_norm": 0.29659849405288696, + "learning_rate": 0.00016615191031800502, + "loss": 0.2768, + "step": 32550 + }, + { + "epoch": 0.5078929306794784, + "grad_norm": 1.5246435403823853, + "learning_rate": 0.00016614151119985027, + "loss": 0.216, + "step": 32560 + }, + { + "epoch": 0.5080489174518001, + "grad_norm": 0.6026854515075684, + "learning_rate": 0.00016613111208169546, + "loss": 0.2277, + "step": 32570 + }, + { + "epoch": 0.5082049042241218, + "grad_norm": 0.41994890570640564, + "learning_rate": 0.0001661207129635407, + "loss": 0.2433, + "step": 32580 + }, + { + "epoch": 0.5083608909964435, + "grad_norm": 2.084599733352661, + "learning_rate": 0.00016611031384538593, + "loss": 0.4081, + "step": 32590 + }, + { + "epoch": 0.5085168777687652, + "grad_norm": 0.2945229709148407, + "learning_rate": 0.00016609991472723115, + "loss": 0.1329, + "step": 32600 + }, + { + "epoch": 0.508672864541087, + "grad_norm": 1.4981423616409302, + "learning_rate": 0.00016608951560907636, + "loss": 0.2855, + "step": 32610 + }, + { + "epoch": 0.5088288513134086, + "grad_norm": 0.961737871170044, + "learning_rate": 0.00016607911649092158, + "loss": 0.1654, + "step": 32620 + }, + { + "epoch": 0.5089848380857304, + "grad_norm": 1.49734628200531, + "learning_rate": 0.0001660687173727668, + "loss": 0.2081, + "step": 32630 + }, + { + "epoch": 0.509140824858052, + "grad_norm": 2.538135051727295, + "learning_rate": 0.00016605831825461202, + "loss": 0.3023, + "step": 32640 + }, + { + "epoch": 0.5092968116303738, + "grad_norm": 0.2981272041797638, + "learning_rate": 0.00016604791913645724, + "loss": 0.214, + "step": 32650 + }, + { + "epoch": 0.5094527984026954, + "grad_norm": 1.8863650560379028, + "learning_rate": 0.00016603752001830246, + "loss": 0.3363, + "step": 32660 + }, + { + "epoch": 0.5096087851750172, + "grad_norm": 0.022720983251929283, + "learning_rate": 0.00016602712090014768, + "loss": 0.2463, + "step": 32670 + }, + { + "epoch": 0.5097647719473388, + "grad_norm": 1.154977798461914, + "learning_rate": 0.0001660167217819929, + "loss": 0.4484, + "step": 32680 + }, + { + "epoch": 0.5099207587196606, + "grad_norm": 0.9470499157905579, + "learning_rate": 0.00016600632266383812, + "loss": 0.1196, + "step": 32690 + }, + { + "epoch": 0.5100767454919822, + "grad_norm": 0.19955401122570038, + "learning_rate": 0.00016599592354568334, + "loss": 0.1762, + "step": 32700 + }, + { + "epoch": 0.510232732264304, + "grad_norm": 0.4098282754421234, + "learning_rate": 0.00016598552442752856, + "loss": 0.2739, + "step": 32710 + }, + { + "epoch": 0.5103887190366257, + "grad_norm": 0.21697258949279785, + "learning_rate": 0.00016597512530937378, + "loss": 0.2555, + "step": 32720 + }, + { + "epoch": 0.5105447058089474, + "grad_norm": 4.894479751586914, + "learning_rate": 0.000165964726191219, + "loss": 0.3137, + "step": 32730 + }, + { + "epoch": 0.5107006925812692, + "grad_norm": 0.8396711349487305, + "learning_rate": 0.00016595432707306422, + "loss": 0.3105, + "step": 32740 + }, + { + "epoch": 0.5108566793535908, + "grad_norm": 1.5494674444198608, + "learning_rate": 0.00016594392795490944, + "loss": 0.2271, + "step": 32750 + }, + { + "epoch": 0.5110126661259126, + "grad_norm": 1.3409888744354248, + "learning_rate": 0.00016593352883675466, + "loss": 0.1796, + "step": 32760 + }, + { + "epoch": 0.5111686528982342, + "grad_norm": 1.7590911388397217, + "learning_rate": 0.00016592312971859988, + "loss": 0.2185, + "step": 32770 + }, + { + "epoch": 0.511324639670556, + "grad_norm": 0.7350002527236938, + "learning_rate": 0.0001659127306004451, + "loss": 0.1911, + "step": 32780 + }, + { + "epoch": 0.5114806264428776, + "grad_norm": 0.7176411747932434, + "learning_rate": 0.00016590233148229032, + "loss": 0.6918, + "step": 32790 + }, + { + "epoch": 0.5116366132151994, + "grad_norm": 2.5005321502685547, + "learning_rate": 0.00016589193236413554, + "loss": 0.2329, + "step": 32800 + }, + { + "epoch": 0.511792599987521, + "grad_norm": 1.3916250467300415, + "learning_rate": 0.00016588153324598076, + "loss": 0.2102, + "step": 32810 + }, + { + "epoch": 0.5119485867598428, + "grad_norm": 0.721671998500824, + "learning_rate": 0.00016587113412782598, + "loss": 0.2965, + "step": 32820 + }, + { + "epoch": 0.5121045735321644, + "grad_norm": 0.6120114326477051, + "learning_rate": 0.0001658607350096712, + "loss": 0.2661, + "step": 32830 + }, + { + "epoch": 0.5122605603044862, + "grad_norm": 0.29123347997665405, + "learning_rate": 0.00016585033589151642, + "loss": 0.2455, + "step": 32840 + }, + { + "epoch": 0.5124165470768078, + "grad_norm": 2.7300868034362793, + "learning_rate": 0.00016583993677336164, + "loss": 0.2049, + "step": 32850 + }, + { + "epoch": 0.5125725338491296, + "grad_norm": 1.801928997039795, + "learning_rate": 0.00016582953765520686, + "loss": 0.1725, + "step": 32860 + }, + { + "epoch": 0.5127285206214514, + "grad_norm": 0.6406471729278564, + "learning_rate": 0.00016581913853705208, + "loss": 0.164, + "step": 32870 + }, + { + "epoch": 0.512884507393773, + "grad_norm": 0.9627325534820557, + "learning_rate": 0.0001658087394188973, + "loss": 0.1728, + "step": 32880 + }, + { + "epoch": 0.5130404941660948, + "grad_norm": 1.5446962118148804, + "learning_rate": 0.00016579834030074251, + "loss": 0.2345, + "step": 32890 + }, + { + "epoch": 0.5131964809384164, + "grad_norm": 0.28092706203460693, + "learning_rate": 0.0001657879411825877, + "loss": 0.2124, + "step": 32900 + }, + { + "epoch": 0.5133524677107382, + "grad_norm": 1.7751421928405762, + "learning_rate": 0.00016577754206443295, + "loss": 0.2367, + "step": 32910 + }, + { + "epoch": 0.5135084544830598, + "grad_norm": 1.6964634656906128, + "learning_rate": 0.00016576714294627815, + "loss": 0.2361, + "step": 32920 + }, + { + "epoch": 0.5136644412553816, + "grad_norm": 0.06759823113679886, + "learning_rate": 0.0001657567438281234, + "loss": 0.2329, + "step": 32930 + }, + { + "epoch": 0.5138204280277032, + "grad_norm": 2.2519280910491943, + "learning_rate": 0.00016574634470996859, + "loss": 0.1079, + "step": 32940 + }, + { + "epoch": 0.513976414800025, + "grad_norm": 3.167576551437378, + "learning_rate": 0.00016573594559181383, + "loss": 0.1764, + "step": 32950 + }, + { + "epoch": 0.5141324015723466, + "grad_norm": 1.6734397411346436, + "learning_rate": 0.00016572554647365903, + "loss": 0.2878, + "step": 32960 + }, + { + "epoch": 0.5142883883446684, + "grad_norm": 0.35627374053001404, + "learning_rate": 0.00016571514735550427, + "loss": 0.2337, + "step": 32970 + }, + { + "epoch": 0.51444437511699, + "grad_norm": 1.6321039199829102, + "learning_rate": 0.00016570474823734946, + "loss": 0.2551, + "step": 32980 + }, + { + "epoch": 0.5146003618893118, + "grad_norm": 1.352982997894287, + "learning_rate": 0.0001656943491191947, + "loss": 0.4467, + "step": 32990 + }, + { + "epoch": 0.5147563486616334, + "grad_norm": 0.3558129072189331, + "learning_rate": 0.0001656839500010399, + "loss": 0.1913, + "step": 33000 + }, + { + "epoch": 0.5149123354339552, + "grad_norm": 0.041972286999225616, + "learning_rate": 0.00016567355088288515, + "loss": 0.1704, + "step": 33010 + }, + { + "epoch": 0.515068322206277, + "grad_norm": 1.4168939590454102, + "learning_rate": 0.00016566315176473034, + "loss": 0.4866, + "step": 33020 + }, + { + "epoch": 0.5152243089785986, + "grad_norm": 3.0042929649353027, + "learning_rate": 0.0001656527526465756, + "loss": 0.2473, + "step": 33030 + }, + { + "epoch": 0.5153802957509204, + "grad_norm": 1.027274489402771, + "learning_rate": 0.00016564235352842078, + "loss": 0.436, + "step": 33040 + }, + { + "epoch": 0.515536282523242, + "grad_norm": 1.2330195903778076, + "learning_rate": 0.00016563195441026603, + "loss": 0.4615, + "step": 33050 + }, + { + "epoch": 0.5156922692955638, + "grad_norm": 2.7145066261291504, + "learning_rate": 0.00016562155529211122, + "loss": 0.278, + "step": 33060 + }, + { + "epoch": 0.5158482560678854, + "grad_norm": 1.1273893117904663, + "learning_rate": 0.00016561115617395647, + "loss": 0.3767, + "step": 33070 + }, + { + "epoch": 0.5160042428402072, + "grad_norm": 2.466265916824341, + "learning_rate": 0.00016560075705580166, + "loss": 0.1879, + "step": 33080 + }, + { + "epoch": 0.5161602296125288, + "grad_norm": 0.3759515583515167, + "learning_rate": 0.0001655903579376469, + "loss": 0.4004, + "step": 33090 + }, + { + "epoch": 0.5163162163848506, + "grad_norm": 1.8821336030960083, + "learning_rate": 0.0001655799588194921, + "loss": 0.1622, + "step": 33100 + }, + { + "epoch": 0.5164722031571722, + "grad_norm": 0.8162096738815308, + "learning_rate": 0.00016556955970133735, + "loss": 0.1341, + "step": 33110 + }, + { + "epoch": 0.516628189929494, + "grad_norm": 0.008498801849782467, + "learning_rate": 0.00016555916058318254, + "loss": 0.2546, + "step": 33120 + }, + { + "epoch": 0.5167841767018156, + "grad_norm": 0.17536675930023193, + "learning_rate": 0.00016554876146502779, + "loss": 0.2934, + "step": 33130 + }, + { + "epoch": 0.5169401634741374, + "grad_norm": 3.539898633956909, + "learning_rate": 0.00016553836234687298, + "loss": 0.4959, + "step": 33140 + }, + { + "epoch": 0.517096150246459, + "grad_norm": 2.190887212753296, + "learning_rate": 0.00016552796322871823, + "loss": 0.2914, + "step": 33150 + }, + { + "epoch": 0.5172521370187808, + "grad_norm": 1.6388943195343018, + "learning_rate": 0.00016551756411056342, + "loss": 0.1317, + "step": 33160 + }, + { + "epoch": 0.5174081237911026, + "grad_norm": 0.33063387870788574, + "learning_rate": 0.00016550716499240866, + "loss": 0.4258, + "step": 33170 + }, + { + "epoch": 0.5175641105634242, + "grad_norm": 0.5168288946151733, + "learning_rate": 0.00016549676587425386, + "loss": 0.2548, + "step": 33180 + }, + { + "epoch": 0.517720097335746, + "grad_norm": 3.9570281505584717, + "learning_rate": 0.0001654863667560991, + "loss": 0.3179, + "step": 33190 + }, + { + "epoch": 0.5178760841080676, + "grad_norm": 1.127255916595459, + "learning_rate": 0.0001654759676379443, + "loss": 0.233, + "step": 33200 + }, + { + "epoch": 0.5180320708803894, + "grad_norm": 0.36299410462379456, + "learning_rate": 0.00016546556851978954, + "loss": 0.4614, + "step": 33210 + }, + { + "epoch": 0.518188057652711, + "grad_norm": 2.914668321609497, + "learning_rate": 0.00016545516940163474, + "loss": 0.1603, + "step": 33220 + }, + { + "epoch": 0.5183440444250328, + "grad_norm": 6.991094589233398, + "learning_rate": 0.00016544477028347998, + "loss": 0.2259, + "step": 33230 + }, + { + "epoch": 0.5185000311973544, + "grad_norm": 2.0594241619110107, + "learning_rate": 0.00016543437116532517, + "loss": 0.1885, + "step": 33240 + }, + { + "epoch": 0.5186560179696762, + "grad_norm": 0.5594693422317505, + "learning_rate": 0.00016542397204717042, + "loss": 0.4333, + "step": 33250 + }, + { + "epoch": 0.5188120047419978, + "grad_norm": 4.314854621887207, + "learning_rate": 0.00016541357292901561, + "loss": 0.2103, + "step": 33260 + }, + { + "epoch": 0.5189679915143196, + "grad_norm": 0.5609000325202942, + "learning_rate": 0.00016540317381086086, + "loss": 0.2498, + "step": 33270 + }, + { + "epoch": 0.5191239782866413, + "grad_norm": 3.680088758468628, + "learning_rate": 0.00016539277469270605, + "loss": 0.3465, + "step": 33280 + }, + { + "epoch": 0.519279965058963, + "grad_norm": 1.4993672370910645, + "learning_rate": 0.0001653823755745513, + "loss": 0.3272, + "step": 33290 + }, + { + "epoch": 0.5194359518312847, + "grad_norm": 0.18005084991455078, + "learning_rate": 0.0001653719764563965, + "loss": 0.1706, + "step": 33300 + }, + { + "epoch": 0.5195919386036064, + "grad_norm": 0.5105524659156799, + "learning_rate": 0.00016536157733824174, + "loss": 0.0937, + "step": 33310 + }, + { + "epoch": 0.5197479253759281, + "grad_norm": 1.2806589603424072, + "learning_rate": 0.00016535117822008693, + "loss": 0.1578, + "step": 33320 + }, + { + "epoch": 0.5199039121482498, + "grad_norm": 0.22015990316867828, + "learning_rate": 0.00016534077910193218, + "loss": 0.2081, + "step": 33330 + }, + { + "epoch": 0.5200598989205716, + "grad_norm": 0.34638991951942444, + "learning_rate": 0.00016533037998377737, + "loss": 0.2389, + "step": 33340 + }, + { + "epoch": 0.5202158856928932, + "grad_norm": 0.8709302544593811, + "learning_rate": 0.0001653199808656226, + "loss": 0.2112, + "step": 33350 + }, + { + "epoch": 0.520371872465215, + "grad_norm": 1.2309521436691284, + "learning_rate": 0.0001653095817474678, + "loss": 0.3448, + "step": 33360 + }, + { + "epoch": 0.5205278592375366, + "grad_norm": 1.3155115842819214, + "learning_rate": 0.00016529918262931303, + "loss": 0.1572, + "step": 33370 + }, + { + "epoch": 0.5206838460098584, + "grad_norm": 0.358751118183136, + "learning_rate": 0.00016528878351115825, + "loss": 0.3011, + "step": 33380 + }, + { + "epoch": 0.52083983278218, + "grad_norm": 3.800628662109375, + "learning_rate": 0.00016527838439300347, + "loss": 0.458, + "step": 33390 + }, + { + "epoch": 0.5209958195545018, + "grad_norm": 1.1253124475479126, + "learning_rate": 0.0001652679852748487, + "loss": 0.1238, + "step": 33400 + }, + { + "epoch": 0.5211518063268235, + "grad_norm": 0.46979108452796936, + "learning_rate": 0.0001652575861566939, + "loss": 0.0745, + "step": 33410 + }, + { + "epoch": 0.5213077930991452, + "grad_norm": 7.921492576599121, + "learning_rate": 0.00016524718703853913, + "loss": 0.4004, + "step": 33420 + }, + { + "epoch": 0.5214637798714669, + "grad_norm": 1.643127679824829, + "learning_rate": 0.00016523678792038435, + "loss": 0.2858, + "step": 33430 + }, + { + "epoch": 0.5216197666437886, + "grad_norm": 7.096631050109863, + "learning_rate": 0.0001652263888022296, + "loss": 0.4735, + "step": 33440 + }, + { + "epoch": 0.5217757534161103, + "grad_norm": 0.7280219197273254, + "learning_rate": 0.0001652159896840748, + "loss": 0.1841, + "step": 33450 + }, + { + "epoch": 0.521931740188432, + "grad_norm": 0.9415522813796997, + "learning_rate": 0.00016520559056592003, + "loss": 0.2169, + "step": 33460 + }, + { + "epoch": 0.5220877269607537, + "grad_norm": 2.264862060546875, + "learning_rate": 0.00016519519144776523, + "loss": 0.1902, + "step": 33470 + }, + { + "epoch": 0.5222437137330754, + "grad_norm": 2.4566822052001953, + "learning_rate": 0.00016518479232961047, + "loss": 0.3329, + "step": 33480 + }, + { + "epoch": 0.5223997005053972, + "grad_norm": 0.7330378890037537, + "learning_rate": 0.00016517439321145567, + "loss": 0.2341, + "step": 33490 + }, + { + "epoch": 0.5225556872777188, + "grad_norm": 1.6409542560577393, + "learning_rate": 0.0001651639940933009, + "loss": 0.3077, + "step": 33500 + }, + { + "epoch": 0.5227116740500406, + "grad_norm": 1.0647432804107666, + "learning_rate": 0.0001651535949751461, + "loss": 0.1935, + "step": 33510 + }, + { + "epoch": 0.5228676608223622, + "grad_norm": 0.507420539855957, + "learning_rate": 0.00016514319585699135, + "loss": 0.1314, + "step": 33520 + }, + { + "epoch": 0.523023647594684, + "grad_norm": 1.0064164400100708, + "learning_rate": 0.00016513279673883654, + "loss": 0.344, + "step": 33530 + }, + { + "epoch": 0.5231796343670057, + "grad_norm": 1.1936209201812744, + "learning_rate": 0.0001651223976206818, + "loss": 0.1948, + "step": 33540 + }, + { + "epoch": 0.5233356211393274, + "grad_norm": 0.04992926865816116, + "learning_rate": 0.00016511199850252698, + "loss": 0.1363, + "step": 33550 + }, + { + "epoch": 0.5234916079116491, + "grad_norm": 1.8256275653839111, + "learning_rate": 0.00016510159938437223, + "loss": 0.443, + "step": 33560 + }, + { + "epoch": 0.5236475946839708, + "grad_norm": 4.914908409118652, + "learning_rate": 0.00016509120026621742, + "loss": 0.4654, + "step": 33570 + }, + { + "epoch": 0.5238035814562925, + "grad_norm": 1.1625486612319946, + "learning_rate": 0.00016508080114806267, + "loss": 0.1974, + "step": 33580 + }, + { + "epoch": 0.5239595682286142, + "grad_norm": 6.6192522048950195, + "learning_rate": 0.00016507040202990786, + "loss": 0.2131, + "step": 33590 + }, + { + "epoch": 0.5241155550009359, + "grad_norm": 0.7369065880775452, + "learning_rate": 0.0001650600029117531, + "loss": 0.2371, + "step": 33600 + }, + { + "epoch": 0.5242715417732576, + "grad_norm": 1.5238152742385864, + "learning_rate": 0.0001650496037935983, + "loss": 0.1557, + "step": 33610 + }, + { + "epoch": 0.5244275285455793, + "grad_norm": 1.0418007373809814, + "learning_rate": 0.00016503920467544355, + "loss": 0.1878, + "step": 33620 + }, + { + "epoch": 0.524583515317901, + "grad_norm": 0.790117084980011, + "learning_rate": 0.00016502880555728874, + "loss": 0.2195, + "step": 33630 + }, + { + "epoch": 0.5247395020902228, + "grad_norm": 1.6712257862091064, + "learning_rate": 0.000165018406439134, + "loss": 0.1602, + "step": 33640 + }, + { + "epoch": 0.5248954888625444, + "grad_norm": 0.19236230850219727, + "learning_rate": 0.00016500800732097918, + "loss": 0.2526, + "step": 33650 + }, + { + "epoch": 0.5250514756348662, + "grad_norm": 1.3519701957702637, + "learning_rate": 0.00016499760820282443, + "loss": 0.4686, + "step": 33660 + }, + { + "epoch": 0.5252074624071879, + "grad_norm": 1.694342851638794, + "learning_rate": 0.00016498720908466962, + "loss": 0.1859, + "step": 33670 + }, + { + "epoch": 0.5253634491795096, + "grad_norm": 5.225239276885986, + "learning_rate": 0.00016497680996651487, + "loss": 0.2944, + "step": 33680 + }, + { + "epoch": 0.5255194359518313, + "grad_norm": 2.0208842754364014, + "learning_rate": 0.00016496641084836006, + "loss": 0.2421, + "step": 33690 + }, + { + "epoch": 0.525675422724153, + "grad_norm": 0.7954996228218079, + "learning_rate": 0.0001649560117302053, + "loss": 0.2853, + "step": 33700 + }, + { + "epoch": 0.5258314094964747, + "grad_norm": 2.296086072921753, + "learning_rate": 0.0001649456126120505, + "loss": 0.1333, + "step": 33710 + }, + { + "epoch": 0.5259873962687964, + "grad_norm": 1.1779128313064575, + "learning_rate": 0.00016493521349389574, + "loss": 0.1066, + "step": 33720 + }, + { + "epoch": 0.5261433830411181, + "grad_norm": 0.1756065934896469, + "learning_rate": 0.00016492481437574094, + "loss": 0.1352, + "step": 33730 + }, + { + "epoch": 0.5262993698134398, + "grad_norm": 0.13100725412368774, + "learning_rate": 0.00016491441525758618, + "loss": 0.2399, + "step": 33740 + }, + { + "epoch": 0.5264553565857615, + "grad_norm": 5.532008171081543, + "learning_rate": 0.00016490401613943138, + "loss": 0.2896, + "step": 33750 + }, + { + "epoch": 0.5266113433580832, + "grad_norm": 1.319886565208435, + "learning_rate": 0.00016489361702127662, + "loss": 0.3275, + "step": 33760 + }, + { + "epoch": 0.5267673301304049, + "grad_norm": 1.5550974607467651, + "learning_rate": 0.00016488321790312182, + "loss": 0.2677, + "step": 33770 + }, + { + "epoch": 0.5269233169027266, + "grad_norm": 1.8936737775802612, + "learning_rate": 0.00016487281878496706, + "loss": 0.1955, + "step": 33780 + }, + { + "epoch": 0.5270793036750484, + "grad_norm": 0.3653401732444763, + "learning_rate": 0.00016486241966681226, + "loss": 0.0723, + "step": 33790 + }, + { + "epoch": 0.52723529044737, + "grad_norm": 2.861341714859009, + "learning_rate": 0.00016485202054865747, + "loss": 0.2412, + "step": 33800 + }, + { + "epoch": 0.5273912772196918, + "grad_norm": 1.5291428565979004, + "learning_rate": 0.0001648416214305027, + "loss": 0.0871, + "step": 33810 + }, + { + "epoch": 0.5275472639920135, + "grad_norm": 1.0372581481933594, + "learning_rate": 0.00016483122231234791, + "loss": 0.4705, + "step": 33820 + }, + { + "epoch": 0.5277032507643352, + "grad_norm": 1.1943141222000122, + "learning_rate": 0.00016482082319419313, + "loss": 0.2848, + "step": 33830 + }, + { + "epoch": 0.5278592375366569, + "grad_norm": 1.9008225202560425, + "learning_rate": 0.00016481042407603835, + "loss": 0.139, + "step": 33840 + }, + { + "epoch": 0.5280152243089786, + "grad_norm": 2.132089138031006, + "learning_rate": 0.00016480002495788357, + "loss": 0.2119, + "step": 33850 + }, + { + "epoch": 0.5281712110813003, + "grad_norm": 0.24524426460266113, + "learning_rate": 0.0001647896258397288, + "loss": 0.1643, + "step": 33860 + }, + { + "epoch": 0.528327197853622, + "grad_norm": 1.6469637155532837, + "learning_rate": 0.000164779226721574, + "loss": 0.363, + "step": 33870 + }, + { + "epoch": 0.5284831846259437, + "grad_norm": 0.8767328858375549, + "learning_rate": 0.00016476882760341923, + "loss": 0.2632, + "step": 33880 + }, + { + "epoch": 0.5286391713982654, + "grad_norm": 0.06347586214542389, + "learning_rate": 0.00016475842848526445, + "loss": 0.3204, + "step": 33890 + }, + { + "epoch": 0.5287951581705871, + "grad_norm": 0.09782540798187256, + "learning_rate": 0.00016474802936710967, + "loss": 0.2211, + "step": 33900 + }, + { + "epoch": 0.5289511449429088, + "grad_norm": 3.2998859882354736, + "learning_rate": 0.0001647376302489549, + "loss": 0.265, + "step": 33910 + }, + { + "epoch": 0.5291071317152305, + "grad_norm": 0.43594226241111755, + "learning_rate": 0.0001647272311308001, + "loss": 0.194, + "step": 33920 + }, + { + "epoch": 0.5292631184875523, + "grad_norm": 1.5166605710983276, + "learning_rate": 0.00016471683201264533, + "loss": 0.2675, + "step": 33930 + }, + { + "epoch": 0.529419105259874, + "grad_norm": 0.6056640148162842, + "learning_rate": 0.00016470643289449055, + "loss": 0.0576, + "step": 33940 + }, + { + "epoch": 0.5295750920321957, + "grad_norm": 0.25410348176956177, + "learning_rate": 0.00016469603377633577, + "loss": 0.1124, + "step": 33950 + }, + { + "epoch": 0.5297310788045174, + "grad_norm": 1.770642876625061, + "learning_rate": 0.000164685634658181, + "loss": 0.3295, + "step": 33960 + }, + { + "epoch": 0.5298870655768391, + "grad_norm": 0.0607205331325531, + "learning_rate": 0.0001646752355400262, + "loss": 0.2369, + "step": 33970 + }, + { + "epoch": 0.5300430523491608, + "grad_norm": 0.5557095408439636, + "learning_rate": 0.00016466483642187143, + "loss": 0.1681, + "step": 33980 + }, + { + "epoch": 0.5301990391214825, + "grad_norm": 0.5192957520484924, + "learning_rate": 0.00016465443730371665, + "loss": 0.2649, + "step": 33990 + }, + { + "epoch": 0.5303550258938042, + "grad_norm": 0.04804835096001625, + "learning_rate": 0.00016464403818556187, + "loss": 0.2639, + "step": 34000 + }, + { + "epoch": 0.5305110126661259, + "grad_norm": 0.02673129364848137, + "learning_rate": 0.0001646336390674071, + "loss": 0.169, + "step": 34010 + }, + { + "epoch": 0.5306669994384476, + "grad_norm": 1.0084244012832642, + "learning_rate": 0.0001646232399492523, + "loss": 0.353, + "step": 34020 + }, + { + "epoch": 0.5308229862107693, + "grad_norm": 2.2202091217041016, + "learning_rate": 0.00016461284083109753, + "loss": 0.3206, + "step": 34030 + }, + { + "epoch": 0.530978972983091, + "grad_norm": 0.5573744773864746, + "learning_rate": 0.00016460244171294275, + "loss": 0.231, + "step": 34040 + }, + { + "epoch": 0.5311349597554127, + "grad_norm": 0.39700084924697876, + "learning_rate": 0.00016459204259478797, + "loss": 0.1412, + "step": 34050 + }, + { + "epoch": 0.5312909465277345, + "grad_norm": 2.582963228225708, + "learning_rate": 0.00016458164347663319, + "loss": 0.3334, + "step": 34060 + }, + { + "epoch": 0.5314469333000561, + "grad_norm": 0.22781169414520264, + "learning_rate": 0.0001645712443584784, + "loss": 0.2767, + "step": 34070 + }, + { + "epoch": 0.5316029200723779, + "grad_norm": 1.2051042318344116, + "learning_rate": 0.00016456084524032362, + "loss": 0.2045, + "step": 34080 + }, + { + "epoch": 0.5317589068446996, + "grad_norm": 0.42760100960731506, + "learning_rate": 0.00016455044612216884, + "loss": 0.1151, + "step": 34090 + }, + { + "epoch": 0.5319148936170213, + "grad_norm": 0.14440476894378662, + "learning_rate": 0.00016454004700401406, + "loss": 0.2386, + "step": 34100 + }, + { + "epoch": 0.532070880389343, + "grad_norm": 2.2777981758117676, + "learning_rate": 0.00016452964788585928, + "loss": 0.2009, + "step": 34110 + }, + { + "epoch": 0.5322268671616647, + "grad_norm": 0.9206979274749756, + "learning_rate": 0.0001645192487677045, + "loss": 0.2745, + "step": 34120 + }, + { + "epoch": 0.5323828539339864, + "grad_norm": 1.6947574615478516, + "learning_rate": 0.00016450884964954972, + "loss": 0.2584, + "step": 34130 + }, + { + "epoch": 0.5325388407063081, + "grad_norm": 0.401444673538208, + "learning_rate": 0.00016449845053139494, + "loss": 0.2218, + "step": 34140 + }, + { + "epoch": 0.5326948274786298, + "grad_norm": 0.08261553198099136, + "learning_rate": 0.00016448805141324016, + "loss": 0.2775, + "step": 34150 + }, + { + "epoch": 0.5328508142509515, + "grad_norm": 0.1017974391579628, + "learning_rate": 0.00016447765229508538, + "loss": 0.2095, + "step": 34160 + }, + { + "epoch": 0.5330068010232732, + "grad_norm": 1.3759571313858032, + "learning_rate": 0.0001644672531769306, + "loss": 0.2643, + "step": 34170 + }, + { + "epoch": 0.5331627877955949, + "grad_norm": 1.2654389142990112, + "learning_rate": 0.00016445685405877582, + "loss": 0.2949, + "step": 34180 + }, + { + "epoch": 0.5333187745679167, + "grad_norm": 1.5481843948364258, + "learning_rate": 0.00016444645494062104, + "loss": 0.1264, + "step": 34190 + }, + { + "epoch": 0.5334747613402383, + "grad_norm": 1.8094528913497925, + "learning_rate": 0.00016443605582246626, + "loss": 0.2727, + "step": 34200 + }, + { + "epoch": 0.5336307481125601, + "grad_norm": 2.224538564682007, + "learning_rate": 0.00016442565670431148, + "loss": 0.3096, + "step": 34210 + }, + { + "epoch": 0.5337867348848817, + "grad_norm": 0.6375226974487305, + "learning_rate": 0.0001644152575861567, + "loss": 0.2251, + "step": 34220 + }, + { + "epoch": 0.5339427216572035, + "grad_norm": 3.727106809616089, + "learning_rate": 0.00016440485846800192, + "loss": 0.4374, + "step": 34230 + }, + { + "epoch": 0.5340987084295252, + "grad_norm": 0.13345426321029663, + "learning_rate": 0.00016439445934984714, + "loss": 0.2011, + "step": 34240 + }, + { + "epoch": 0.5342546952018469, + "grad_norm": 2.1658668518066406, + "learning_rate": 0.00016438406023169236, + "loss": 0.2457, + "step": 34250 + }, + { + "epoch": 0.5344106819741686, + "grad_norm": 19.238407135009766, + "learning_rate": 0.00016437366111353758, + "loss": 0.2756, + "step": 34260 + }, + { + "epoch": 0.5345666687464903, + "grad_norm": 1.0292778015136719, + "learning_rate": 0.0001643632619953828, + "loss": 0.1646, + "step": 34270 + }, + { + "epoch": 0.534722655518812, + "grad_norm": 0.9372987747192383, + "learning_rate": 0.00016435286287722802, + "loss": 0.2762, + "step": 34280 + }, + { + "epoch": 0.5348786422911337, + "grad_norm": 0.3918002247810364, + "learning_rate": 0.00016434246375907324, + "loss": 0.1406, + "step": 34290 + }, + { + "epoch": 0.5350346290634554, + "grad_norm": 1.3518732786178589, + "learning_rate": 0.00016433206464091846, + "loss": 0.2992, + "step": 34300 + }, + { + "epoch": 0.5351906158357771, + "grad_norm": 0.73117595911026, + "learning_rate": 0.00016432166552276368, + "loss": 0.1172, + "step": 34310 + }, + { + "epoch": 0.5353466026080989, + "grad_norm": 0.06655958294868469, + "learning_rate": 0.0001643112664046089, + "loss": 0.1332, + "step": 34320 + }, + { + "epoch": 0.5355025893804205, + "grad_norm": 3.2587168216705322, + "learning_rate": 0.00016430086728645412, + "loss": 0.2052, + "step": 34330 + }, + { + "epoch": 0.5356585761527423, + "grad_norm": 2.6668431758880615, + "learning_rate": 0.00016429046816829934, + "loss": 0.1295, + "step": 34340 + }, + { + "epoch": 0.5358145629250639, + "grad_norm": 1.316080927848816, + "learning_rate": 0.00016428006905014456, + "loss": 0.2072, + "step": 34350 + }, + { + "epoch": 0.5359705496973857, + "grad_norm": 5.144528388977051, + "learning_rate": 0.00016426966993198977, + "loss": 0.1934, + "step": 34360 + }, + { + "epoch": 0.5361265364697073, + "grad_norm": 1.3529599905014038, + "learning_rate": 0.000164259270813835, + "loss": 0.2077, + "step": 34370 + }, + { + "epoch": 0.5362825232420291, + "grad_norm": 0.9132925868034363, + "learning_rate": 0.00016424887169568021, + "loss": 0.2404, + "step": 34380 + }, + { + "epoch": 0.5364385100143508, + "grad_norm": 1.520033597946167, + "learning_rate": 0.00016423847257752543, + "loss": 0.1765, + "step": 34390 + }, + { + "epoch": 0.5365944967866725, + "grad_norm": 2.343975305557251, + "learning_rate": 0.00016422807345937065, + "loss": 0.1959, + "step": 34400 + }, + { + "epoch": 0.5367504835589942, + "grad_norm": 0.5855118632316589, + "learning_rate": 0.00016421767434121587, + "loss": 0.0848, + "step": 34410 + }, + { + "epoch": 0.5369064703313159, + "grad_norm": 1.9108648300170898, + "learning_rate": 0.0001642072752230611, + "loss": 0.2406, + "step": 34420 + }, + { + "epoch": 0.5370624571036376, + "grad_norm": 1.0323792695999146, + "learning_rate": 0.0001641968761049063, + "loss": 0.2298, + "step": 34430 + }, + { + "epoch": 0.5372184438759593, + "grad_norm": 0.17435620725154877, + "learning_rate": 0.00016418647698675153, + "loss": 0.1396, + "step": 34440 + }, + { + "epoch": 0.537374430648281, + "grad_norm": 0.3526577353477478, + "learning_rate": 0.00016417607786859675, + "loss": 0.172, + "step": 34450 + }, + { + "epoch": 0.5375304174206027, + "grad_norm": 2.73685884475708, + "learning_rate": 0.00016416567875044197, + "loss": 0.1167, + "step": 34460 + }, + { + "epoch": 0.5376864041929245, + "grad_norm": 2.765693426132202, + "learning_rate": 0.0001641552796322872, + "loss": 0.1356, + "step": 34470 + }, + { + "epoch": 0.5378423909652461, + "grad_norm": 0.47628021240234375, + "learning_rate": 0.0001641448805141324, + "loss": 0.3104, + "step": 34480 + }, + { + "epoch": 0.5379983777375679, + "grad_norm": 1.1307590007781982, + "learning_rate": 0.00016413448139597763, + "loss": 0.3247, + "step": 34490 + }, + { + "epoch": 0.5381543645098895, + "grad_norm": 0.07136381417512894, + "learning_rate": 0.00016412408227782285, + "loss": 0.319, + "step": 34500 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.100904722432e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-34500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..52bffcef594a2a740999baa79533435532c5e63f --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eddb2067e3d1106a977829b35b6793fc2ce7bdee39cdcadfe0778e12d7c74c3a +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..f0544486acf7cc3c9190501cdea0a21eef44b9e7 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d46b4c0f2f1c344f0a1f5424bb7b7e21bc4d9f183feb5882729f55c6d0ba11b +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..7f8eab436ab05d49d6413e362cfa5bb7b9fbce1a Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..292da2babdd7e53906400fd0a9f4220a1c89efaf Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..d4b8d54629308c19f4ffd390434c9349b3e6ddf0 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/trainer_state.json @@ -0,0 +1,24534 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5459537031259749, + "eval_steps": 500, + "global_step": 35000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + }, + { + "epoch": 0.5071129968178698, + "grad_norm": 1.0075199604034424, + "learning_rate": 0.00016619350679062414, + "loss": 0.2569, + "step": 32510 + }, + { + "epoch": 0.5072689835901916, + "grad_norm": 1.3489066362380981, + "learning_rate": 0.0001661831076724694, + "loss": 0.2565, + "step": 32520 + }, + { + "epoch": 0.5074249703625132, + "grad_norm": 0.2534058094024658, + "learning_rate": 0.00016617270855431458, + "loss": 0.2292, + "step": 32530 + }, + { + "epoch": 0.507580957134835, + "grad_norm": 0.31723281741142273, + "learning_rate": 0.00016616230943615983, + "loss": 0.1467, + "step": 32540 + }, + { + "epoch": 0.5077369439071566, + "grad_norm": 0.29659849405288696, + "learning_rate": 0.00016615191031800502, + "loss": 0.2768, + "step": 32550 + }, + { + "epoch": 0.5078929306794784, + "grad_norm": 1.5246435403823853, + "learning_rate": 0.00016614151119985027, + "loss": 0.216, + "step": 32560 + }, + { + "epoch": 0.5080489174518001, + "grad_norm": 0.6026854515075684, + "learning_rate": 0.00016613111208169546, + "loss": 0.2277, + "step": 32570 + }, + { + "epoch": 0.5082049042241218, + "grad_norm": 0.41994890570640564, + "learning_rate": 0.0001661207129635407, + "loss": 0.2433, + "step": 32580 + }, + { + "epoch": 0.5083608909964435, + "grad_norm": 2.084599733352661, + "learning_rate": 0.00016611031384538593, + "loss": 0.4081, + "step": 32590 + }, + { + "epoch": 0.5085168777687652, + "grad_norm": 0.2945229709148407, + "learning_rate": 0.00016609991472723115, + "loss": 0.1329, + "step": 32600 + }, + { + "epoch": 0.508672864541087, + "grad_norm": 1.4981423616409302, + "learning_rate": 0.00016608951560907636, + "loss": 0.2855, + "step": 32610 + }, + { + "epoch": 0.5088288513134086, + "grad_norm": 0.961737871170044, + "learning_rate": 0.00016607911649092158, + "loss": 0.1654, + "step": 32620 + }, + { + "epoch": 0.5089848380857304, + "grad_norm": 1.49734628200531, + "learning_rate": 0.0001660687173727668, + "loss": 0.2081, + "step": 32630 + }, + { + "epoch": 0.509140824858052, + "grad_norm": 2.538135051727295, + "learning_rate": 0.00016605831825461202, + "loss": 0.3023, + "step": 32640 + }, + { + "epoch": 0.5092968116303738, + "grad_norm": 0.2981272041797638, + "learning_rate": 0.00016604791913645724, + "loss": 0.214, + "step": 32650 + }, + { + "epoch": 0.5094527984026954, + "grad_norm": 1.8863650560379028, + "learning_rate": 0.00016603752001830246, + "loss": 0.3363, + "step": 32660 + }, + { + "epoch": 0.5096087851750172, + "grad_norm": 0.022720983251929283, + "learning_rate": 0.00016602712090014768, + "loss": 0.2463, + "step": 32670 + }, + { + "epoch": 0.5097647719473388, + "grad_norm": 1.154977798461914, + "learning_rate": 0.0001660167217819929, + "loss": 0.4484, + "step": 32680 + }, + { + "epoch": 0.5099207587196606, + "grad_norm": 0.9470499157905579, + "learning_rate": 0.00016600632266383812, + "loss": 0.1196, + "step": 32690 + }, + { + "epoch": 0.5100767454919822, + "grad_norm": 0.19955401122570038, + "learning_rate": 0.00016599592354568334, + "loss": 0.1762, + "step": 32700 + }, + { + "epoch": 0.510232732264304, + "grad_norm": 0.4098282754421234, + "learning_rate": 0.00016598552442752856, + "loss": 0.2739, + "step": 32710 + }, + { + "epoch": 0.5103887190366257, + "grad_norm": 0.21697258949279785, + "learning_rate": 0.00016597512530937378, + "loss": 0.2555, + "step": 32720 + }, + { + "epoch": 0.5105447058089474, + "grad_norm": 4.894479751586914, + "learning_rate": 0.000165964726191219, + "loss": 0.3137, + "step": 32730 + }, + { + "epoch": 0.5107006925812692, + "grad_norm": 0.8396711349487305, + "learning_rate": 0.00016595432707306422, + "loss": 0.3105, + "step": 32740 + }, + { + "epoch": 0.5108566793535908, + "grad_norm": 1.5494674444198608, + "learning_rate": 0.00016594392795490944, + "loss": 0.2271, + "step": 32750 + }, + { + "epoch": 0.5110126661259126, + "grad_norm": 1.3409888744354248, + "learning_rate": 0.00016593352883675466, + "loss": 0.1796, + "step": 32760 + }, + { + "epoch": 0.5111686528982342, + "grad_norm": 1.7590911388397217, + "learning_rate": 0.00016592312971859988, + "loss": 0.2185, + "step": 32770 + }, + { + "epoch": 0.511324639670556, + "grad_norm": 0.7350002527236938, + "learning_rate": 0.0001659127306004451, + "loss": 0.1911, + "step": 32780 + }, + { + "epoch": 0.5114806264428776, + "grad_norm": 0.7176411747932434, + "learning_rate": 0.00016590233148229032, + "loss": 0.6918, + "step": 32790 + }, + { + "epoch": 0.5116366132151994, + "grad_norm": 2.5005321502685547, + "learning_rate": 0.00016589193236413554, + "loss": 0.2329, + "step": 32800 + }, + { + "epoch": 0.511792599987521, + "grad_norm": 1.3916250467300415, + "learning_rate": 0.00016588153324598076, + "loss": 0.2102, + "step": 32810 + }, + { + "epoch": 0.5119485867598428, + "grad_norm": 0.721671998500824, + "learning_rate": 0.00016587113412782598, + "loss": 0.2965, + "step": 32820 + }, + { + "epoch": 0.5121045735321644, + "grad_norm": 0.6120114326477051, + "learning_rate": 0.0001658607350096712, + "loss": 0.2661, + "step": 32830 + }, + { + "epoch": 0.5122605603044862, + "grad_norm": 0.29123347997665405, + "learning_rate": 0.00016585033589151642, + "loss": 0.2455, + "step": 32840 + }, + { + "epoch": 0.5124165470768078, + "grad_norm": 2.7300868034362793, + "learning_rate": 0.00016583993677336164, + "loss": 0.2049, + "step": 32850 + }, + { + "epoch": 0.5125725338491296, + "grad_norm": 1.801928997039795, + "learning_rate": 0.00016582953765520686, + "loss": 0.1725, + "step": 32860 + }, + { + "epoch": 0.5127285206214514, + "grad_norm": 0.6406471729278564, + "learning_rate": 0.00016581913853705208, + "loss": 0.164, + "step": 32870 + }, + { + "epoch": 0.512884507393773, + "grad_norm": 0.9627325534820557, + "learning_rate": 0.0001658087394188973, + "loss": 0.1728, + "step": 32880 + }, + { + "epoch": 0.5130404941660948, + "grad_norm": 1.5446962118148804, + "learning_rate": 0.00016579834030074251, + "loss": 0.2345, + "step": 32890 + }, + { + "epoch": 0.5131964809384164, + "grad_norm": 0.28092706203460693, + "learning_rate": 0.0001657879411825877, + "loss": 0.2124, + "step": 32900 + }, + { + "epoch": 0.5133524677107382, + "grad_norm": 1.7751421928405762, + "learning_rate": 0.00016577754206443295, + "loss": 0.2367, + "step": 32910 + }, + { + "epoch": 0.5135084544830598, + "grad_norm": 1.6964634656906128, + "learning_rate": 0.00016576714294627815, + "loss": 0.2361, + "step": 32920 + }, + { + "epoch": 0.5136644412553816, + "grad_norm": 0.06759823113679886, + "learning_rate": 0.0001657567438281234, + "loss": 0.2329, + "step": 32930 + }, + { + "epoch": 0.5138204280277032, + "grad_norm": 2.2519280910491943, + "learning_rate": 0.00016574634470996859, + "loss": 0.1079, + "step": 32940 + }, + { + "epoch": 0.513976414800025, + "grad_norm": 3.167576551437378, + "learning_rate": 0.00016573594559181383, + "loss": 0.1764, + "step": 32950 + }, + { + "epoch": 0.5141324015723466, + "grad_norm": 1.6734397411346436, + "learning_rate": 0.00016572554647365903, + "loss": 0.2878, + "step": 32960 + }, + { + "epoch": 0.5142883883446684, + "grad_norm": 0.35627374053001404, + "learning_rate": 0.00016571514735550427, + "loss": 0.2337, + "step": 32970 + }, + { + "epoch": 0.51444437511699, + "grad_norm": 1.6321039199829102, + "learning_rate": 0.00016570474823734946, + "loss": 0.2551, + "step": 32980 + }, + { + "epoch": 0.5146003618893118, + "grad_norm": 1.352982997894287, + "learning_rate": 0.0001656943491191947, + "loss": 0.4467, + "step": 32990 + }, + { + "epoch": 0.5147563486616334, + "grad_norm": 0.3558129072189331, + "learning_rate": 0.0001656839500010399, + "loss": 0.1913, + "step": 33000 + }, + { + "epoch": 0.5149123354339552, + "grad_norm": 0.041972286999225616, + "learning_rate": 0.00016567355088288515, + "loss": 0.1704, + "step": 33010 + }, + { + "epoch": 0.515068322206277, + "grad_norm": 1.4168939590454102, + "learning_rate": 0.00016566315176473034, + "loss": 0.4866, + "step": 33020 + }, + { + "epoch": 0.5152243089785986, + "grad_norm": 3.0042929649353027, + "learning_rate": 0.0001656527526465756, + "loss": 0.2473, + "step": 33030 + }, + { + "epoch": 0.5153802957509204, + "grad_norm": 1.027274489402771, + "learning_rate": 0.00016564235352842078, + "loss": 0.436, + "step": 33040 + }, + { + "epoch": 0.515536282523242, + "grad_norm": 1.2330195903778076, + "learning_rate": 0.00016563195441026603, + "loss": 0.4615, + "step": 33050 + }, + { + "epoch": 0.5156922692955638, + "grad_norm": 2.7145066261291504, + "learning_rate": 0.00016562155529211122, + "loss": 0.278, + "step": 33060 + }, + { + "epoch": 0.5158482560678854, + "grad_norm": 1.1273893117904663, + "learning_rate": 0.00016561115617395647, + "loss": 0.3767, + "step": 33070 + }, + { + "epoch": 0.5160042428402072, + "grad_norm": 2.466265916824341, + "learning_rate": 0.00016560075705580166, + "loss": 0.1879, + "step": 33080 + }, + { + "epoch": 0.5161602296125288, + "grad_norm": 0.3759515583515167, + "learning_rate": 0.0001655903579376469, + "loss": 0.4004, + "step": 33090 + }, + { + "epoch": 0.5163162163848506, + "grad_norm": 1.8821336030960083, + "learning_rate": 0.0001655799588194921, + "loss": 0.1622, + "step": 33100 + }, + { + "epoch": 0.5164722031571722, + "grad_norm": 0.8162096738815308, + "learning_rate": 0.00016556955970133735, + "loss": 0.1341, + "step": 33110 + }, + { + "epoch": 0.516628189929494, + "grad_norm": 0.008498801849782467, + "learning_rate": 0.00016555916058318254, + "loss": 0.2546, + "step": 33120 + }, + { + "epoch": 0.5167841767018156, + "grad_norm": 0.17536675930023193, + "learning_rate": 0.00016554876146502779, + "loss": 0.2934, + "step": 33130 + }, + { + "epoch": 0.5169401634741374, + "grad_norm": 3.539898633956909, + "learning_rate": 0.00016553836234687298, + "loss": 0.4959, + "step": 33140 + }, + { + "epoch": 0.517096150246459, + "grad_norm": 2.190887212753296, + "learning_rate": 0.00016552796322871823, + "loss": 0.2914, + "step": 33150 + }, + { + "epoch": 0.5172521370187808, + "grad_norm": 1.6388943195343018, + "learning_rate": 0.00016551756411056342, + "loss": 0.1317, + "step": 33160 + }, + { + "epoch": 0.5174081237911026, + "grad_norm": 0.33063387870788574, + "learning_rate": 0.00016550716499240866, + "loss": 0.4258, + "step": 33170 + }, + { + "epoch": 0.5175641105634242, + "grad_norm": 0.5168288946151733, + "learning_rate": 0.00016549676587425386, + "loss": 0.2548, + "step": 33180 + }, + { + "epoch": 0.517720097335746, + "grad_norm": 3.9570281505584717, + "learning_rate": 0.0001654863667560991, + "loss": 0.3179, + "step": 33190 + }, + { + "epoch": 0.5178760841080676, + "grad_norm": 1.127255916595459, + "learning_rate": 0.0001654759676379443, + "loss": 0.233, + "step": 33200 + }, + { + "epoch": 0.5180320708803894, + "grad_norm": 0.36299410462379456, + "learning_rate": 0.00016546556851978954, + "loss": 0.4614, + "step": 33210 + }, + { + "epoch": 0.518188057652711, + "grad_norm": 2.914668321609497, + "learning_rate": 0.00016545516940163474, + "loss": 0.1603, + "step": 33220 + }, + { + "epoch": 0.5183440444250328, + "grad_norm": 6.991094589233398, + "learning_rate": 0.00016544477028347998, + "loss": 0.2259, + "step": 33230 + }, + { + "epoch": 0.5185000311973544, + "grad_norm": 2.0594241619110107, + "learning_rate": 0.00016543437116532517, + "loss": 0.1885, + "step": 33240 + }, + { + "epoch": 0.5186560179696762, + "grad_norm": 0.5594693422317505, + "learning_rate": 0.00016542397204717042, + "loss": 0.4333, + "step": 33250 + }, + { + "epoch": 0.5188120047419978, + "grad_norm": 4.314854621887207, + "learning_rate": 0.00016541357292901561, + "loss": 0.2103, + "step": 33260 + }, + { + "epoch": 0.5189679915143196, + "grad_norm": 0.5609000325202942, + "learning_rate": 0.00016540317381086086, + "loss": 0.2498, + "step": 33270 + }, + { + "epoch": 0.5191239782866413, + "grad_norm": 3.680088758468628, + "learning_rate": 0.00016539277469270605, + "loss": 0.3465, + "step": 33280 + }, + { + "epoch": 0.519279965058963, + "grad_norm": 1.4993672370910645, + "learning_rate": 0.0001653823755745513, + "loss": 0.3272, + "step": 33290 + }, + { + "epoch": 0.5194359518312847, + "grad_norm": 0.18005084991455078, + "learning_rate": 0.0001653719764563965, + "loss": 0.1706, + "step": 33300 + }, + { + "epoch": 0.5195919386036064, + "grad_norm": 0.5105524659156799, + "learning_rate": 0.00016536157733824174, + "loss": 0.0937, + "step": 33310 + }, + { + "epoch": 0.5197479253759281, + "grad_norm": 1.2806589603424072, + "learning_rate": 0.00016535117822008693, + "loss": 0.1578, + "step": 33320 + }, + { + "epoch": 0.5199039121482498, + "grad_norm": 0.22015990316867828, + "learning_rate": 0.00016534077910193218, + "loss": 0.2081, + "step": 33330 + }, + { + "epoch": 0.5200598989205716, + "grad_norm": 0.34638991951942444, + "learning_rate": 0.00016533037998377737, + "loss": 0.2389, + "step": 33340 + }, + { + "epoch": 0.5202158856928932, + "grad_norm": 0.8709302544593811, + "learning_rate": 0.0001653199808656226, + "loss": 0.2112, + "step": 33350 + }, + { + "epoch": 0.520371872465215, + "grad_norm": 1.2309521436691284, + "learning_rate": 0.0001653095817474678, + "loss": 0.3448, + "step": 33360 + }, + { + "epoch": 0.5205278592375366, + "grad_norm": 1.3155115842819214, + "learning_rate": 0.00016529918262931303, + "loss": 0.1572, + "step": 33370 + }, + { + "epoch": 0.5206838460098584, + "grad_norm": 0.358751118183136, + "learning_rate": 0.00016528878351115825, + "loss": 0.3011, + "step": 33380 + }, + { + "epoch": 0.52083983278218, + "grad_norm": 3.800628662109375, + "learning_rate": 0.00016527838439300347, + "loss": 0.458, + "step": 33390 + }, + { + "epoch": 0.5209958195545018, + "grad_norm": 1.1253124475479126, + "learning_rate": 0.0001652679852748487, + "loss": 0.1238, + "step": 33400 + }, + { + "epoch": 0.5211518063268235, + "grad_norm": 0.46979108452796936, + "learning_rate": 0.0001652575861566939, + "loss": 0.0745, + "step": 33410 + }, + { + "epoch": 0.5213077930991452, + "grad_norm": 7.921492576599121, + "learning_rate": 0.00016524718703853913, + "loss": 0.4004, + "step": 33420 + }, + { + "epoch": 0.5214637798714669, + "grad_norm": 1.643127679824829, + "learning_rate": 0.00016523678792038435, + "loss": 0.2858, + "step": 33430 + }, + { + "epoch": 0.5216197666437886, + "grad_norm": 7.096631050109863, + "learning_rate": 0.0001652263888022296, + "loss": 0.4735, + "step": 33440 + }, + { + "epoch": 0.5217757534161103, + "grad_norm": 0.7280219197273254, + "learning_rate": 0.0001652159896840748, + "loss": 0.1841, + "step": 33450 + }, + { + "epoch": 0.521931740188432, + "grad_norm": 0.9415522813796997, + "learning_rate": 0.00016520559056592003, + "loss": 0.2169, + "step": 33460 + }, + { + "epoch": 0.5220877269607537, + "grad_norm": 2.264862060546875, + "learning_rate": 0.00016519519144776523, + "loss": 0.1902, + "step": 33470 + }, + { + "epoch": 0.5222437137330754, + "grad_norm": 2.4566822052001953, + "learning_rate": 0.00016518479232961047, + "loss": 0.3329, + "step": 33480 + }, + { + "epoch": 0.5223997005053972, + "grad_norm": 0.7330378890037537, + "learning_rate": 0.00016517439321145567, + "loss": 0.2341, + "step": 33490 + }, + { + "epoch": 0.5225556872777188, + "grad_norm": 1.6409542560577393, + "learning_rate": 0.0001651639940933009, + "loss": 0.3077, + "step": 33500 + }, + { + "epoch": 0.5227116740500406, + "grad_norm": 1.0647432804107666, + "learning_rate": 0.0001651535949751461, + "loss": 0.1935, + "step": 33510 + }, + { + "epoch": 0.5228676608223622, + "grad_norm": 0.507420539855957, + "learning_rate": 0.00016514319585699135, + "loss": 0.1314, + "step": 33520 + }, + { + "epoch": 0.523023647594684, + "grad_norm": 1.0064164400100708, + "learning_rate": 0.00016513279673883654, + "loss": 0.344, + "step": 33530 + }, + { + "epoch": 0.5231796343670057, + "grad_norm": 1.1936209201812744, + "learning_rate": 0.0001651223976206818, + "loss": 0.1948, + "step": 33540 + }, + { + "epoch": 0.5233356211393274, + "grad_norm": 0.04992926865816116, + "learning_rate": 0.00016511199850252698, + "loss": 0.1363, + "step": 33550 + }, + { + "epoch": 0.5234916079116491, + "grad_norm": 1.8256275653839111, + "learning_rate": 0.00016510159938437223, + "loss": 0.443, + "step": 33560 + }, + { + "epoch": 0.5236475946839708, + "grad_norm": 4.914908409118652, + "learning_rate": 0.00016509120026621742, + "loss": 0.4654, + "step": 33570 + }, + { + "epoch": 0.5238035814562925, + "grad_norm": 1.1625486612319946, + "learning_rate": 0.00016508080114806267, + "loss": 0.1974, + "step": 33580 + }, + { + "epoch": 0.5239595682286142, + "grad_norm": 6.6192522048950195, + "learning_rate": 0.00016507040202990786, + "loss": 0.2131, + "step": 33590 + }, + { + "epoch": 0.5241155550009359, + "grad_norm": 0.7369065880775452, + "learning_rate": 0.0001650600029117531, + "loss": 0.2371, + "step": 33600 + }, + { + "epoch": 0.5242715417732576, + "grad_norm": 1.5238152742385864, + "learning_rate": 0.0001650496037935983, + "loss": 0.1557, + "step": 33610 + }, + { + "epoch": 0.5244275285455793, + "grad_norm": 1.0418007373809814, + "learning_rate": 0.00016503920467544355, + "loss": 0.1878, + "step": 33620 + }, + { + "epoch": 0.524583515317901, + "grad_norm": 0.790117084980011, + "learning_rate": 0.00016502880555728874, + "loss": 0.2195, + "step": 33630 + }, + { + "epoch": 0.5247395020902228, + "grad_norm": 1.6712257862091064, + "learning_rate": 0.000165018406439134, + "loss": 0.1602, + "step": 33640 + }, + { + "epoch": 0.5248954888625444, + "grad_norm": 0.19236230850219727, + "learning_rate": 0.00016500800732097918, + "loss": 0.2526, + "step": 33650 + }, + { + "epoch": 0.5250514756348662, + "grad_norm": 1.3519701957702637, + "learning_rate": 0.00016499760820282443, + "loss": 0.4686, + "step": 33660 + }, + { + "epoch": 0.5252074624071879, + "grad_norm": 1.694342851638794, + "learning_rate": 0.00016498720908466962, + "loss": 0.1859, + "step": 33670 + }, + { + "epoch": 0.5253634491795096, + "grad_norm": 5.225239276885986, + "learning_rate": 0.00016497680996651487, + "loss": 0.2944, + "step": 33680 + }, + { + "epoch": 0.5255194359518313, + "grad_norm": 2.0208842754364014, + "learning_rate": 0.00016496641084836006, + "loss": 0.2421, + "step": 33690 + }, + { + "epoch": 0.525675422724153, + "grad_norm": 0.7954996228218079, + "learning_rate": 0.0001649560117302053, + "loss": 0.2853, + "step": 33700 + }, + { + "epoch": 0.5258314094964747, + "grad_norm": 2.296086072921753, + "learning_rate": 0.0001649456126120505, + "loss": 0.1333, + "step": 33710 + }, + { + "epoch": 0.5259873962687964, + "grad_norm": 1.1779128313064575, + "learning_rate": 0.00016493521349389574, + "loss": 0.1066, + "step": 33720 + }, + { + "epoch": 0.5261433830411181, + "grad_norm": 0.1756065934896469, + "learning_rate": 0.00016492481437574094, + "loss": 0.1352, + "step": 33730 + }, + { + "epoch": 0.5262993698134398, + "grad_norm": 0.13100725412368774, + "learning_rate": 0.00016491441525758618, + "loss": 0.2399, + "step": 33740 + }, + { + "epoch": 0.5264553565857615, + "grad_norm": 5.532008171081543, + "learning_rate": 0.00016490401613943138, + "loss": 0.2896, + "step": 33750 + }, + { + "epoch": 0.5266113433580832, + "grad_norm": 1.319886565208435, + "learning_rate": 0.00016489361702127662, + "loss": 0.3275, + "step": 33760 + }, + { + "epoch": 0.5267673301304049, + "grad_norm": 1.5550974607467651, + "learning_rate": 0.00016488321790312182, + "loss": 0.2677, + "step": 33770 + }, + { + "epoch": 0.5269233169027266, + "grad_norm": 1.8936737775802612, + "learning_rate": 0.00016487281878496706, + "loss": 0.1955, + "step": 33780 + }, + { + "epoch": 0.5270793036750484, + "grad_norm": 0.3653401732444763, + "learning_rate": 0.00016486241966681226, + "loss": 0.0723, + "step": 33790 + }, + { + "epoch": 0.52723529044737, + "grad_norm": 2.861341714859009, + "learning_rate": 0.00016485202054865747, + "loss": 0.2412, + "step": 33800 + }, + { + "epoch": 0.5273912772196918, + "grad_norm": 1.5291428565979004, + "learning_rate": 0.0001648416214305027, + "loss": 0.0871, + "step": 33810 + }, + { + "epoch": 0.5275472639920135, + "grad_norm": 1.0372581481933594, + "learning_rate": 0.00016483122231234791, + "loss": 0.4705, + "step": 33820 + }, + { + "epoch": 0.5277032507643352, + "grad_norm": 1.1943141222000122, + "learning_rate": 0.00016482082319419313, + "loss": 0.2848, + "step": 33830 + }, + { + "epoch": 0.5278592375366569, + "grad_norm": 1.9008225202560425, + "learning_rate": 0.00016481042407603835, + "loss": 0.139, + "step": 33840 + }, + { + "epoch": 0.5280152243089786, + "grad_norm": 2.132089138031006, + "learning_rate": 0.00016480002495788357, + "loss": 0.2119, + "step": 33850 + }, + { + "epoch": 0.5281712110813003, + "grad_norm": 0.24524426460266113, + "learning_rate": 0.0001647896258397288, + "loss": 0.1643, + "step": 33860 + }, + { + "epoch": 0.528327197853622, + "grad_norm": 1.6469637155532837, + "learning_rate": 0.000164779226721574, + "loss": 0.363, + "step": 33870 + }, + { + "epoch": 0.5284831846259437, + "grad_norm": 0.8767328858375549, + "learning_rate": 0.00016476882760341923, + "loss": 0.2632, + "step": 33880 + }, + { + "epoch": 0.5286391713982654, + "grad_norm": 0.06347586214542389, + "learning_rate": 0.00016475842848526445, + "loss": 0.3204, + "step": 33890 + }, + { + "epoch": 0.5287951581705871, + "grad_norm": 0.09782540798187256, + "learning_rate": 0.00016474802936710967, + "loss": 0.2211, + "step": 33900 + }, + { + "epoch": 0.5289511449429088, + "grad_norm": 3.2998859882354736, + "learning_rate": 0.0001647376302489549, + "loss": 0.265, + "step": 33910 + }, + { + "epoch": 0.5291071317152305, + "grad_norm": 0.43594226241111755, + "learning_rate": 0.0001647272311308001, + "loss": 0.194, + "step": 33920 + }, + { + "epoch": 0.5292631184875523, + "grad_norm": 1.5166605710983276, + "learning_rate": 0.00016471683201264533, + "loss": 0.2675, + "step": 33930 + }, + { + "epoch": 0.529419105259874, + "grad_norm": 0.6056640148162842, + "learning_rate": 0.00016470643289449055, + "loss": 0.0576, + "step": 33940 + }, + { + "epoch": 0.5295750920321957, + "grad_norm": 0.25410348176956177, + "learning_rate": 0.00016469603377633577, + "loss": 0.1124, + "step": 33950 + }, + { + "epoch": 0.5297310788045174, + "grad_norm": 1.770642876625061, + "learning_rate": 0.000164685634658181, + "loss": 0.3295, + "step": 33960 + }, + { + "epoch": 0.5298870655768391, + "grad_norm": 0.0607205331325531, + "learning_rate": 0.0001646752355400262, + "loss": 0.2369, + "step": 33970 + }, + { + "epoch": 0.5300430523491608, + "grad_norm": 0.5557095408439636, + "learning_rate": 0.00016466483642187143, + "loss": 0.1681, + "step": 33980 + }, + { + "epoch": 0.5301990391214825, + "grad_norm": 0.5192957520484924, + "learning_rate": 0.00016465443730371665, + "loss": 0.2649, + "step": 33990 + }, + { + "epoch": 0.5303550258938042, + "grad_norm": 0.04804835096001625, + "learning_rate": 0.00016464403818556187, + "loss": 0.2639, + "step": 34000 + }, + { + "epoch": 0.5305110126661259, + "grad_norm": 0.02673129364848137, + "learning_rate": 0.0001646336390674071, + "loss": 0.169, + "step": 34010 + }, + { + "epoch": 0.5306669994384476, + "grad_norm": 1.0084244012832642, + "learning_rate": 0.0001646232399492523, + "loss": 0.353, + "step": 34020 + }, + { + "epoch": 0.5308229862107693, + "grad_norm": 2.2202091217041016, + "learning_rate": 0.00016461284083109753, + "loss": 0.3206, + "step": 34030 + }, + { + "epoch": 0.530978972983091, + "grad_norm": 0.5573744773864746, + "learning_rate": 0.00016460244171294275, + "loss": 0.231, + "step": 34040 + }, + { + "epoch": 0.5311349597554127, + "grad_norm": 0.39700084924697876, + "learning_rate": 0.00016459204259478797, + "loss": 0.1412, + "step": 34050 + }, + { + "epoch": 0.5312909465277345, + "grad_norm": 2.582963228225708, + "learning_rate": 0.00016458164347663319, + "loss": 0.3334, + "step": 34060 + }, + { + "epoch": 0.5314469333000561, + "grad_norm": 0.22781169414520264, + "learning_rate": 0.0001645712443584784, + "loss": 0.2767, + "step": 34070 + }, + { + "epoch": 0.5316029200723779, + "grad_norm": 1.2051042318344116, + "learning_rate": 0.00016456084524032362, + "loss": 0.2045, + "step": 34080 + }, + { + "epoch": 0.5317589068446996, + "grad_norm": 0.42760100960731506, + "learning_rate": 0.00016455044612216884, + "loss": 0.1151, + "step": 34090 + }, + { + "epoch": 0.5319148936170213, + "grad_norm": 0.14440476894378662, + "learning_rate": 0.00016454004700401406, + "loss": 0.2386, + "step": 34100 + }, + { + "epoch": 0.532070880389343, + "grad_norm": 2.2777981758117676, + "learning_rate": 0.00016452964788585928, + "loss": 0.2009, + "step": 34110 + }, + { + "epoch": 0.5322268671616647, + "grad_norm": 0.9206979274749756, + "learning_rate": 0.0001645192487677045, + "loss": 0.2745, + "step": 34120 + }, + { + "epoch": 0.5323828539339864, + "grad_norm": 1.6947574615478516, + "learning_rate": 0.00016450884964954972, + "loss": 0.2584, + "step": 34130 + }, + { + "epoch": 0.5325388407063081, + "grad_norm": 0.401444673538208, + "learning_rate": 0.00016449845053139494, + "loss": 0.2218, + "step": 34140 + }, + { + "epoch": 0.5326948274786298, + "grad_norm": 0.08261553198099136, + "learning_rate": 0.00016448805141324016, + "loss": 0.2775, + "step": 34150 + }, + { + "epoch": 0.5328508142509515, + "grad_norm": 0.1017974391579628, + "learning_rate": 0.00016447765229508538, + "loss": 0.2095, + "step": 34160 + }, + { + "epoch": 0.5330068010232732, + "grad_norm": 1.3759571313858032, + "learning_rate": 0.0001644672531769306, + "loss": 0.2643, + "step": 34170 + }, + { + "epoch": 0.5331627877955949, + "grad_norm": 1.2654389142990112, + "learning_rate": 0.00016445685405877582, + "loss": 0.2949, + "step": 34180 + }, + { + "epoch": 0.5333187745679167, + "grad_norm": 1.5481843948364258, + "learning_rate": 0.00016444645494062104, + "loss": 0.1264, + "step": 34190 + }, + { + "epoch": 0.5334747613402383, + "grad_norm": 1.8094528913497925, + "learning_rate": 0.00016443605582246626, + "loss": 0.2727, + "step": 34200 + }, + { + "epoch": 0.5336307481125601, + "grad_norm": 2.224538564682007, + "learning_rate": 0.00016442565670431148, + "loss": 0.3096, + "step": 34210 + }, + { + "epoch": 0.5337867348848817, + "grad_norm": 0.6375226974487305, + "learning_rate": 0.0001644152575861567, + "loss": 0.2251, + "step": 34220 + }, + { + "epoch": 0.5339427216572035, + "grad_norm": 3.727106809616089, + "learning_rate": 0.00016440485846800192, + "loss": 0.4374, + "step": 34230 + }, + { + "epoch": 0.5340987084295252, + "grad_norm": 0.13345426321029663, + "learning_rate": 0.00016439445934984714, + "loss": 0.2011, + "step": 34240 + }, + { + "epoch": 0.5342546952018469, + "grad_norm": 2.1658668518066406, + "learning_rate": 0.00016438406023169236, + "loss": 0.2457, + "step": 34250 + }, + { + "epoch": 0.5344106819741686, + "grad_norm": 19.238407135009766, + "learning_rate": 0.00016437366111353758, + "loss": 0.2756, + "step": 34260 + }, + { + "epoch": 0.5345666687464903, + "grad_norm": 1.0292778015136719, + "learning_rate": 0.0001643632619953828, + "loss": 0.1646, + "step": 34270 + }, + { + "epoch": 0.534722655518812, + "grad_norm": 0.9372987747192383, + "learning_rate": 0.00016435286287722802, + "loss": 0.2762, + "step": 34280 + }, + { + "epoch": 0.5348786422911337, + "grad_norm": 0.3918002247810364, + "learning_rate": 0.00016434246375907324, + "loss": 0.1406, + "step": 34290 + }, + { + "epoch": 0.5350346290634554, + "grad_norm": 1.3518732786178589, + "learning_rate": 0.00016433206464091846, + "loss": 0.2992, + "step": 34300 + }, + { + "epoch": 0.5351906158357771, + "grad_norm": 0.73117595911026, + "learning_rate": 0.00016432166552276368, + "loss": 0.1172, + "step": 34310 + }, + { + "epoch": 0.5353466026080989, + "grad_norm": 0.06655958294868469, + "learning_rate": 0.0001643112664046089, + "loss": 0.1332, + "step": 34320 + }, + { + "epoch": 0.5355025893804205, + "grad_norm": 3.2587168216705322, + "learning_rate": 0.00016430086728645412, + "loss": 0.2052, + "step": 34330 + }, + { + "epoch": 0.5356585761527423, + "grad_norm": 2.6668431758880615, + "learning_rate": 0.00016429046816829934, + "loss": 0.1295, + "step": 34340 + }, + { + "epoch": 0.5358145629250639, + "grad_norm": 1.316080927848816, + "learning_rate": 0.00016428006905014456, + "loss": 0.2072, + "step": 34350 + }, + { + "epoch": 0.5359705496973857, + "grad_norm": 5.144528388977051, + "learning_rate": 0.00016426966993198977, + "loss": 0.1934, + "step": 34360 + }, + { + "epoch": 0.5361265364697073, + "grad_norm": 1.3529599905014038, + "learning_rate": 0.000164259270813835, + "loss": 0.2077, + "step": 34370 + }, + { + "epoch": 0.5362825232420291, + "grad_norm": 0.9132925868034363, + "learning_rate": 0.00016424887169568021, + "loss": 0.2404, + "step": 34380 + }, + { + "epoch": 0.5364385100143508, + "grad_norm": 1.520033597946167, + "learning_rate": 0.00016423847257752543, + "loss": 0.1765, + "step": 34390 + }, + { + "epoch": 0.5365944967866725, + "grad_norm": 2.343975305557251, + "learning_rate": 0.00016422807345937065, + "loss": 0.1959, + "step": 34400 + }, + { + "epoch": 0.5367504835589942, + "grad_norm": 0.5855118632316589, + "learning_rate": 0.00016421767434121587, + "loss": 0.0848, + "step": 34410 + }, + { + "epoch": 0.5369064703313159, + "grad_norm": 1.9108648300170898, + "learning_rate": 0.0001642072752230611, + "loss": 0.2406, + "step": 34420 + }, + { + "epoch": 0.5370624571036376, + "grad_norm": 1.0323792695999146, + "learning_rate": 0.0001641968761049063, + "loss": 0.2298, + "step": 34430 + }, + { + "epoch": 0.5372184438759593, + "grad_norm": 0.17435620725154877, + "learning_rate": 0.00016418647698675153, + "loss": 0.1396, + "step": 34440 + }, + { + "epoch": 0.537374430648281, + "grad_norm": 0.3526577353477478, + "learning_rate": 0.00016417607786859675, + "loss": 0.172, + "step": 34450 + }, + { + "epoch": 0.5375304174206027, + "grad_norm": 2.73685884475708, + "learning_rate": 0.00016416567875044197, + "loss": 0.1167, + "step": 34460 + }, + { + "epoch": 0.5376864041929245, + "grad_norm": 2.765693426132202, + "learning_rate": 0.0001641552796322872, + "loss": 0.1356, + "step": 34470 + }, + { + "epoch": 0.5378423909652461, + "grad_norm": 0.47628021240234375, + "learning_rate": 0.0001641448805141324, + "loss": 0.3104, + "step": 34480 + }, + { + "epoch": 0.5379983777375679, + "grad_norm": 1.1307590007781982, + "learning_rate": 0.00016413448139597763, + "loss": 0.3247, + "step": 34490 + }, + { + "epoch": 0.5381543645098895, + "grad_norm": 0.07136381417512894, + "learning_rate": 0.00016412408227782285, + "loss": 0.319, + "step": 34500 + }, + { + "epoch": 0.5383103512822113, + "grad_norm": 1.7850221395492554, + "learning_rate": 0.00016411368315966807, + "loss": 0.2515, + "step": 34510 + }, + { + "epoch": 0.5384663380545329, + "grad_norm": 1.3870742321014404, + "learning_rate": 0.0001641032840415133, + "loss": 0.256, + "step": 34520 + }, + { + "epoch": 0.5386223248268547, + "grad_norm": 1.1329221725463867, + "learning_rate": 0.0001640928849233585, + "loss": 0.2255, + "step": 34530 + }, + { + "epoch": 0.5387783115991764, + "grad_norm": 0.23236137628555298, + "learning_rate": 0.00016408248580520373, + "loss": 0.4053, + "step": 34540 + }, + { + "epoch": 0.5389342983714981, + "grad_norm": 3.350924253463745, + "learning_rate": 0.00016407208668704895, + "loss": 0.2816, + "step": 34550 + }, + { + "epoch": 0.5390902851438198, + "grad_norm": 1.7459170818328857, + "learning_rate": 0.00016406168756889417, + "loss": 0.0834, + "step": 34560 + }, + { + "epoch": 0.5392462719161415, + "grad_norm": 0.89351487159729, + "learning_rate": 0.0001640512884507394, + "loss": 0.1146, + "step": 34570 + }, + { + "epoch": 0.5394022586884633, + "grad_norm": 0.49410831928253174, + "learning_rate": 0.0001640408893325846, + "loss": 0.2319, + "step": 34580 + }, + { + "epoch": 0.5395582454607849, + "grad_norm": 0.009971237741410732, + "learning_rate": 0.00016403049021442983, + "loss": 0.1674, + "step": 34590 + }, + { + "epoch": 0.5397142322331067, + "grad_norm": 1.420511245727539, + "learning_rate": 0.00016402009109627505, + "loss": 0.1469, + "step": 34600 + }, + { + "epoch": 0.5398702190054283, + "grad_norm": 2.5686769485473633, + "learning_rate": 0.00016400969197812027, + "loss": 0.2607, + "step": 34610 + }, + { + "epoch": 0.5400262057777501, + "grad_norm": 2.5808913707733154, + "learning_rate": 0.00016399929285996549, + "loss": 0.4412, + "step": 34620 + }, + { + "epoch": 0.5401821925500717, + "grad_norm": 0.2778591811656952, + "learning_rate": 0.0001639888937418107, + "loss": 0.1563, + "step": 34630 + }, + { + "epoch": 0.5403381793223935, + "grad_norm": 1.6448099613189697, + "learning_rate": 0.00016397849462365592, + "loss": 0.2484, + "step": 34640 + }, + { + "epoch": 0.5404941660947151, + "grad_norm": 0.003213417250663042, + "learning_rate": 0.00016396809550550114, + "loss": 0.1582, + "step": 34650 + }, + { + "epoch": 0.5406501528670369, + "grad_norm": 2.3433990478515625, + "learning_rate": 0.00016395769638734636, + "loss": 0.2774, + "step": 34660 + }, + { + "epoch": 0.5408061396393585, + "grad_norm": 0.09544426947832108, + "learning_rate": 0.00016394729726919158, + "loss": 0.1242, + "step": 34670 + }, + { + "epoch": 0.5409621264116803, + "grad_norm": 1.8934235572814941, + "learning_rate": 0.0001639368981510368, + "loss": 0.1636, + "step": 34680 + }, + { + "epoch": 0.541118113184002, + "grad_norm": 2.8528554439544678, + "learning_rate": 0.00016392649903288202, + "loss": 0.2524, + "step": 34690 + }, + { + "epoch": 0.5412740999563237, + "grad_norm": 2.837836980819702, + "learning_rate": 0.00016391609991472722, + "loss": 0.1808, + "step": 34700 + }, + { + "epoch": 0.5414300867286455, + "grad_norm": 4.349740028381348, + "learning_rate": 0.00016390570079657246, + "loss": 0.12, + "step": 34710 + }, + { + "epoch": 0.5415860735009671, + "grad_norm": 0.6011945605278015, + "learning_rate": 0.00016389530167841765, + "loss": 0.2865, + "step": 34720 + }, + { + "epoch": 0.5417420602732889, + "grad_norm": 1.2079488039016724, + "learning_rate": 0.0001638849025602629, + "loss": 0.125, + "step": 34730 + }, + { + "epoch": 0.5418980470456105, + "grad_norm": 1.9549164772033691, + "learning_rate": 0.0001638745034421081, + "loss": 0.3285, + "step": 34740 + }, + { + "epoch": 0.5420540338179323, + "grad_norm": 0.6969407200813293, + "learning_rate": 0.00016386410432395334, + "loss": 0.3267, + "step": 34750 + }, + { + "epoch": 0.5422100205902539, + "grad_norm": 1.0561965703964233, + "learning_rate": 0.00016385370520579853, + "loss": 0.1463, + "step": 34760 + }, + { + "epoch": 0.5423660073625757, + "grad_norm": 1.8208931684494019, + "learning_rate": 0.00016384330608764378, + "loss": 0.1713, + "step": 34770 + }, + { + "epoch": 0.5425219941348973, + "grad_norm": 0.6349910497665405, + "learning_rate": 0.000163832906969489, + "loss": 0.2244, + "step": 34780 + }, + { + "epoch": 0.5426779809072191, + "grad_norm": 1.1976515054702759, + "learning_rate": 0.00016382250785133422, + "loss": 0.3407, + "step": 34790 + }, + { + "epoch": 0.5428339676795407, + "grad_norm": 0.5970319509506226, + "learning_rate": 0.00016381210873317944, + "loss": 0.1419, + "step": 34800 + }, + { + "epoch": 0.5429899544518625, + "grad_norm": 0.14623159170150757, + "learning_rate": 0.00016380170961502466, + "loss": 0.14, + "step": 34810 + }, + { + "epoch": 0.5431459412241841, + "grad_norm": 1.4332351684570312, + "learning_rate": 0.00016379131049686988, + "loss": 0.1093, + "step": 34820 + }, + { + "epoch": 0.5433019279965059, + "grad_norm": 1.1443063020706177, + "learning_rate": 0.0001637809113787151, + "loss": 0.3676, + "step": 34830 + }, + { + "epoch": 0.5434579147688277, + "grad_norm": 1.3897461891174316, + "learning_rate": 0.00016377051226056032, + "loss": 0.3713, + "step": 34840 + }, + { + "epoch": 0.5436139015411493, + "grad_norm": 1.7773199081420898, + "learning_rate": 0.00016376011314240554, + "loss": 0.2249, + "step": 34850 + }, + { + "epoch": 0.5437698883134711, + "grad_norm": 0.8524389266967773, + "learning_rate": 0.00016374971402425076, + "loss": 0.2158, + "step": 34860 + }, + { + "epoch": 0.5439258750857927, + "grad_norm": 0.5572389960289001, + "learning_rate": 0.00016373931490609598, + "loss": 0.215, + "step": 34870 + }, + { + "epoch": 0.5440818618581145, + "grad_norm": 3.0238709449768066, + "learning_rate": 0.0001637289157879412, + "loss": 0.3765, + "step": 34880 + }, + { + "epoch": 0.5442378486304361, + "grad_norm": 1.1655421257019043, + "learning_rate": 0.00016371851666978642, + "loss": 0.2491, + "step": 34890 + }, + { + "epoch": 0.5443938354027579, + "grad_norm": 0.49492090940475464, + "learning_rate": 0.00016370811755163164, + "loss": 0.1287, + "step": 34900 + }, + { + "epoch": 0.5445498221750795, + "grad_norm": 0.2732921540737152, + "learning_rate": 0.00016369771843347685, + "loss": 0.2154, + "step": 34910 + }, + { + "epoch": 0.5447058089474013, + "grad_norm": 2.5807769298553467, + "learning_rate": 0.00016368731931532207, + "loss": 0.2032, + "step": 34920 + }, + { + "epoch": 0.5448617957197229, + "grad_norm": 0.5044315457344055, + "learning_rate": 0.0001636769201971673, + "loss": 0.2595, + "step": 34930 + }, + { + "epoch": 0.5450177824920447, + "grad_norm": 2.1248972415924072, + "learning_rate": 0.00016366652107901251, + "loss": 0.5651, + "step": 34940 + }, + { + "epoch": 0.5451737692643663, + "grad_norm": 0.8391468524932861, + "learning_rate": 0.00016365612196085773, + "loss": 0.3028, + "step": 34950 + }, + { + "epoch": 0.5453297560366881, + "grad_norm": 0.36081477999687195, + "learning_rate": 0.00016364572284270295, + "loss": 0.1073, + "step": 34960 + }, + { + "epoch": 0.5454857428090097, + "grad_norm": 3.175804376602173, + "learning_rate": 0.00016363532372454817, + "loss": 0.1889, + "step": 34970 + }, + { + "epoch": 0.5456417295813315, + "grad_norm": 0.20647937059402466, + "learning_rate": 0.0001636249246063934, + "loss": 0.2416, + "step": 34980 + }, + { + "epoch": 0.5457977163536533, + "grad_norm": 2.3525943756103516, + "learning_rate": 0.0001636145254882386, + "loss": 0.3179, + "step": 34990 + }, + { + "epoch": 0.5459537031259749, + "grad_norm": 0.2641216814517975, + "learning_rate": 0.00016360412637008383, + "loss": 0.2157, + "step": 35000 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.13135261696e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d13210fd5642352bc778d442fe220bf33bee002c --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b7c31a057d9d2f925e0af01e2d6e17d10dbb06d762ca3fa30fa92644eacae77 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..2ca2047f8150468e20dee3a5d8184e46ae43c88b --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1530500645ca6fc18161ea89350a21ee4f391f99b7f37dd33d7e0c826458fc9 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..1f5300bbdf1bd82f9b76e80c866491320c0e0999 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..943afeb39ba677085aec5506b11b3cc745495494 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..3557bcf545d4b7b21d7b5e42243db393ba94a9fa --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/trainer_state.json @@ -0,0 +1,24884 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5537530417420603, + "eval_steps": 500, + "global_step": 35500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + }, + { + "epoch": 0.5071129968178698, + "grad_norm": 1.0075199604034424, + "learning_rate": 0.00016619350679062414, + "loss": 0.2569, + "step": 32510 + }, + { + "epoch": 0.5072689835901916, + "grad_norm": 1.3489066362380981, + "learning_rate": 0.0001661831076724694, + "loss": 0.2565, + "step": 32520 + }, + { + "epoch": 0.5074249703625132, + "grad_norm": 0.2534058094024658, + "learning_rate": 0.00016617270855431458, + "loss": 0.2292, + "step": 32530 + }, + { + "epoch": 0.507580957134835, + "grad_norm": 0.31723281741142273, + "learning_rate": 0.00016616230943615983, + "loss": 0.1467, + "step": 32540 + }, + { + "epoch": 0.5077369439071566, + "grad_norm": 0.29659849405288696, + "learning_rate": 0.00016615191031800502, + "loss": 0.2768, + "step": 32550 + }, + { + "epoch": 0.5078929306794784, + "grad_norm": 1.5246435403823853, + "learning_rate": 0.00016614151119985027, + "loss": 0.216, + "step": 32560 + }, + { + "epoch": 0.5080489174518001, + "grad_norm": 0.6026854515075684, + "learning_rate": 0.00016613111208169546, + "loss": 0.2277, + "step": 32570 + }, + { + "epoch": 0.5082049042241218, + "grad_norm": 0.41994890570640564, + "learning_rate": 0.0001661207129635407, + "loss": 0.2433, + "step": 32580 + }, + { + "epoch": 0.5083608909964435, + "grad_norm": 2.084599733352661, + "learning_rate": 0.00016611031384538593, + "loss": 0.4081, + "step": 32590 + }, + { + "epoch": 0.5085168777687652, + "grad_norm": 0.2945229709148407, + "learning_rate": 0.00016609991472723115, + "loss": 0.1329, + "step": 32600 + }, + { + "epoch": 0.508672864541087, + "grad_norm": 1.4981423616409302, + "learning_rate": 0.00016608951560907636, + "loss": 0.2855, + "step": 32610 + }, + { + "epoch": 0.5088288513134086, + "grad_norm": 0.961737871170044, + "learning_rate": 0.00016607911649092158, + "loss": 0.1654, + "step": 32620 + }, + { + "epoch": 0.5089848380857304, + "grad_norm": 1.49734628200531, + "learning_rate": 0.0001660687173727668, + "loss": 0.2081, + "step": 32630 + }, + { + "epoch": 0.509140824858052, + "grad_norm": 2.538135051727295, + "learning_rate": 0.00016605831825461202, + "loss": 0.3023, + "step": 32640 + }, + { + "epoch": 0.5092968116303738, + "grad_norm": 0.2981272041797638, + "learning_rate": 0.00016604791913645724, + "loss": 0.214, + "step": 32650 + }, + { + "epoch": 0.5094527984026954, + "grad_norm": 1.8863650560379028, + "learning_rate": 0.00016603752001830246, + "loss": 0.3363, + "step": 32660 + }, + { + "epoch": 0.5096087851750172, + "grad_norm": 0.022720983251929283, + "learning_rate": 0.00016602712090014768, + "loss": 0.2463, + "step": 32670 + }, + { + "epoch": 0.5097647719473388, + "grad_norm": 1.154977798461914, + "learning_rate": 0.0001660167217819929, + "loss": 0.4484, + "step": 32680 + }, + { + "epoch": 0.5099207587196606, + "grad_norm": 0.9470499157905579, + "learning_rate": 0.00016600632266383812, + "loss": 0.1196, + "step": 32690 + }, + { + "epoch": 0.5100767454919822, + "grad_norm": 0.19955401122570038, + "learning_rate": 0.00016599592354568334, + "loss": 0.1762, + "step": 32700 + }, + { + "epoch": 0.510232732264304, + "grad_norm": 0.4098282754421234, + "learning_rate": 0.00016598552442752856, + "loss": 0.2739, + "step": 32710 + }, + { + "epoch": 0.5103887190366257, + "grad_norm": 0.21697258949279785, + "learning_rate": 0.00016597512530937378, + "loss": 0.2555, + "step": 32720 + }, + { + "epoch": 0.5105447058089474, + "grad_norm": 4.894479751586914, + "learning_rate": 0.000165964726191219, + "loss": 0.3137, + "step": 32730 + }, + { + "epoch": 0.5107006925812692, + "grad_norm": 0.8396711349487305, + "learning_rate": 0.00016595432707306422, + "loss": 0.3105, + "step": 32740 + }, + { + "epoch": 0.5108566793535908, + "grad_norm": 1.5494674444198608, + "learning_rate": 0.00016594392795490944, + "loss": 0.2271, + "step": 32750 + }, + { + "epoch": 0.5110126661259126, + "grad_norm": 1.3409888744354248, + "learning_rate": 0.00016593352883675466, + "loss": 0.1796, + "step": 32760 + }, + { + "epoch": 0.5111686528982342, + "grad_norm": 1.7590911388397217, + "learning_rate": 0.00016592312971859988, + "loss": 0.2185, + "step": 32770 + }, + { + "epoch": 0.511324639670556, + "grad_norm": 0.7350002527236938, + "learning_rate": 0.0001659127306004451, + "loss": 0.1911, + "step": 32780 + }, + { + "epoch": 0.5114806264428776, + "grad_norm": 0.7176411747932434, + "learning_rate": 0.00016590233148229032, + "loss": 0.6918, + "step": 32790 + }, + { + "epoch": 0.5116366132151994, + "grad_norm": 2.5005321502685547, + "learning_rate": 0.00016589193236413554, + "loss": 0.2329, + "step": 32800 + }, + { + "epoch": 0.511792599987521, + "grad_norm": 1.3916250467300415, + "learning_rate": 0.00016588153324598076, + "loss": 0.2102, + "step": 32810 + }, + { + "epoch": 0.5119485867598428, + "grad_norm": 0.721671998500824, + "learning_rate": 0.00016587113412782598, + "loss": 0.2965, + "step": 32820 + }, + { + "epoch": 0.5121045735321644, + "grad_norm": 0.6120114326477051, + "learning_rate": 0.0001658607350096712, + "loss": 0.2661, + "step": 32830 + }, + { + "epoch": 0.5122605603044862, + "grad_norm": 0.29123347997665405, + "learning_rate": 0.00016585033589151642, + "loss": 0.2455, + "step": 32840 + }, + { + "epoch": 0.5124165470768078, + "grad_norm": 2.7300868034362793, + "learning_rate": 0.00016583993677336164, + "loss": 0.2049, + "step": 32850 + }, + { + "epoch": 0.5125725338491296, + "grad_norm": 1.801928997039795, + "learning_rate": 0.00016582953765520686, + "loss": 0.1725, + "step": 32860 + }, + { + "epoch": 0.5127285206214514, + "grad_norm": 0.6406471729278564, + "learning_rate": 0.00016581913853705208, + "loss": 0.164, + "step": 32870 + }, + { + "epoch": 0.512884507393773, + "grad_norm": 0.9627325534820557, + "learning_rate": 0.0001658087394188973, + "loss": 0.1728, + "step": 32880 + }, + { + "epoch": 0.5130404941660948, + "grad_norm": 1.5446962118148804, + "learning_rate": 0.00016579834030074251, + "loss": 0.2345, + "step": 32890 + }, + { + "epoch": 0.5131964809384164, + "grad_norm": 0.28092706203460693, + "learning_rate": 0.0001657879411825877, + "loss": 0.2124, + "step": 32900 + }, + { + "epoch": 0.5133524677107382, + "grad_norm": 1.7751421928405762, + "learning_rate": 0.00016577754206443295, + "loss": 0.2367, + "step": 32910 + }, + { + "epoch": 0.5135084544830598, + "grad_norm": 1.6964634656906128, + "learning_rate": 0.00016576714294627815, + "loss": 0.2361, + "step": 32920 + }, + { + "epoch": 0.5136644412553816, + "grad_norm": 0.06759823113679886, + "learning_rate": 0.0001657567438281234, + "loss": 0.2329, + "step": 32930 + }, + { + "epoch": 0.5138204280277032, + "grad_norm": 2.2519280910491943, + "learning_rate": 0.00016574634470996859, + "loss": 0.1079, + "step": 32940 + }, + { + "epoch": 0.513976414800025, + "grad_norm": 3.167576551437378, + "learning_rate": 0.00016573594559181383, + "loss": 0.1764, + "step": 32950 + }, + { + "epoch": 0.5141324015723466, + "grad_norm": 1.6734397411346436, + "learning_rate": 0.00016572554647365903, + "loss": 0.2878, + "step": 32960 + }, + { + "epoch": 0.5142883883446684, + "grad_norm": 0.35627374053001404, + "learning_rate": 0.00016571514735550427, + "loss": 0.2337, + "step": 32970 + }, + { + "epoch": 0.51444437511699, + "grad_norm": 1.6321039199829102, + "learning_rate": 0.00016570474823734946, + "loss": 0.2551, + "step": 32980 + }, + { + "epoch": 0.5146003618893118, + "grad_norm": 1.352982997894287, + "learning_rate": 0.0001656943491191947, + "loss": 0.4467, + "step": 32990 + }, + { + "epoch": 0.5147563486616334, + "grad_norm": 0.3558129072189331, + "learning_rate": 0.0001656839500010399, + "loss": 0.1913, + "step": 33000 + }, + { + "epoch": 0.5149123354339552, + "grad_norm": 0.041972286999225616, + "learning_rate": 0.00016567355088288515, + "loss": 0.1704, + "step": 33010 + }, + { + "epoch": 0.515068322206277, + "grad_norm": 1.4168939590454102, + "learning_rate": 0.00016566315176473034, + "loss": 0.4866, + "step": 33020 + }, + { + "epoch": 0.5152243089785986, + "grad_norm": 3.0042929649353027, + "learning_rate": 0.0001656527526465756, + "loss": 0.2473, + "step": 33030 + }, + { + "epoch": 0.5153802957509204, + "grad_norm": 1.027274489402771, + "learning_rate": 0.00016564235352842078, + "loss": 0.436, + "step": 33040 + }, + { + "epoch": 0.515536282523242, + "grad_norm": 1.2330195903778076, + "learning_rate": 0.00016563195441026603, + "loss": 0.4615, + "step": 33050 + }, + { + "epoch": 0.5156922692955638, + "grad_norm": 2.7145066261291504, + "learning_rate": 0.00016562155529211122, + "loss": 0.278, + "step": 33060 + }, + { + "epoch": 0.5158482560678854, + "grad_norm": 1.1273893117904663, + "learning_rate": 0.00016561115617395647, + "loss": 0.3767, + "step": 33070 + }, + { + "epoch": 0.5160042428402072, + "grad_norm": 2.466265916824341, + "learning_rate": 0.00016560075705580166, + "loss": 0.1879, + "step": 33080 + }, + { + "epoch": 0.5161602296125288, + "grad_norm": 0.3759515583515167, + "learning_rate": 0.0001655903579376469, + "loss": 0.4004, + "step": 33090 + }, + { + "epoch": 0.5163162163848506, + "grad_norm": 1.8821336030960083, + "learning_rate": 0.0001655799588194921, + "loss": 0.1622, + "step": 33100 + }, + { + "epoch": 0.5164722031571722, + "grad_norm": 0.8162096738815308, + "learning_rate": 0.00016556955970133735, + "loss": 0.1341, + "step": 33110 + }, + { + "epoch": 0.516628189929494, + "grad_norm": 0.008498801849782467, + "learning_rate": 0.00016555916058318254, + "loss": 0.2546, + "step": 33120 + }, + { + "epoch": 0.5167841767018156, + "grad_norm": 0.17536675930023193, + "learning_rate": 0.00016554876146502779, + "loss": 0.2934, + "step": 33130 + }, + { + "epoch": 0.5169401634741374, + "grad_norm": 3.539898633956909, + "learning_rate": 0.00016553836234687298, + "loss": 0.4959, + "step": 33140 + }, + { + "epoch": 0.517096150246459, + "grad_norm": 2.190887212753296, + "learning_rate": 0.00016552796322871823, + "loss": 0.2914, + "step": 33150 + }, + { + "epoch": 0.5172521370187808, + "grad_norm": 1.6388943195343018, + "learning_rate": 0.00016551756411056342, + "loss": 0.1317, + "step": 33160 + }, + { + "epoch": 0.5174081237911026, + "grad_norm": 0.33063387870788574, + "learning_rate": 0.00016550716499240866, + "loss": 0.4258, + "step": 33170 + }, + { + "epoch": 0.5175641105634242, + "grad_norm": 0.5168288946151733, + "learning_rate": 0.00016549676587425386, + "loss": 0.2548, + "step": 33180 + }, + { + "epoch": 0.517720097335746, + "grad_norm": 3.9570281505584717, + "learning_rate": 0.0001654863667560991, + "loss": 0.3179, + "step": 33190 + }, + { + "epoch": 0.5178760841080676, + "grad_norm": 1.127255916595459, + "learning_rate": 0.0001654759676379443, + "loss": 0.233, + "step": 33200 + }, + { + "epoch": 0.5180320708803894, + "grad_norm": 0.36299410462379456, + "learning_rate": 0.00016546556851978954, + "loss": 0.4614, + "step": 33210 + }, + { + "epoch": 0.518188057652711, + "grad_norm": 2.914668321609497, + "learning_rate": 0.00016545516940163474, + "loss": 0.1603, + "step": 33220 + }, + { + "epoch": 0.5183440444250328, + "grad_norm": 6.991094589233398, + "learning_rate": 0.00016544477028347998, + "loss": 0.2259, + "step": 33230 + }, + { + "epoch": 0.5185000311973544, + "grad_norm": 2.0594241619110107, + "learning_rate": 0.00016543437116532517, + "loss": 0.1885, + "step": 33240 + }, + { + "epoch": 0.5186560179696762, + "grad_norm": 0.5594693422317505, + "learning_rate": 0.00016542397204717042, + "loss": 0.4333, + "step": 33250 + }, + { + "epoch": 0.5188120047419978, + "grad_norm": 4.314854621887207, + "learning_rate": 0.00016541357292901561, + "loss": 0.2103, + "step": 33260 + }, + { + "epoch": 0.5189679915143196, + "grad_norm": 0.5609000325202942, + "learning_rate": 0.00016540317381086086, + "loss": 0.2498, + "step": 33270 + }, + { + "epoch": 0.5191239782866413, + "grad_norm": 3.680088758468628, + "learning_rate": 0.00016539277469270605, + "loss": 0.3465, + "step": 33280 + }, + { + "epoch": 0.519279965058963, + "grad_norm": 1.4993672370910645, + "learning_rate": 0.0001653823755745513, + "loss": 0.3272, + "step": 33290 + }, + { + "epoch": 0.5194359518312847, + "grad_norm": 0.18005084991455078, + "learning_rate": 0.0001653719764563965, + "loss": 0.1706, + "step": 33300 + }, + { + "epoch": 0.5195919386036064, + "grad_norm": 0.5105524659156799, + "learning_rate": 0.00016536157733824174, + "loss": 0.0937, + "step": 33310 + }, + { + "epoch": 0.5197479253759281, + "grad_norm": 1.2806589603424072, + "learning_rate": 0.00016535117822008693, + "loss": 0.1578, + "step": 33320 + }, + { + "epoch": 0.5199039121482498, + "grad_norm": 0.22015990316867828, + "learning_rate": 0.00016534077910193218, + "loss": 0.2081, + "step": 33330 + }, + { + "epoch": 0.5200598989205716, + "grad_norm": 0.34638991951942444, + "learning_rate": 0.00016533037998377737, + "loss": 0.2389, + "step": 33340 + }, + { + "epoch": 0.5202158856928932, + "grad_norm": 0.8709302544593811, + "learning_rate": 0.0001653199808656226, + "loss": 0.2112, + "step": 33350 + }, + { + "epoch": 0.520371872465215, + "grad_norm": 1.2309521436691284, + "learning_rate": 0.0001653095817474678, + "loss": 0.3448, + "step": 33360 + }, + { + "epoch": 0.5205278592375366, + "grad_norm": 1.3155115842819214, + "learning_rate": 0.00016529918262931303, + "loss": 0.1572, + "step": 33370 + }, + { + "epoch": 0.5206838460098584, + "grad_norm": 0.358751118183136, + "learning_rate": 0.00016528878351115825, + "loss": 0.3011, + "step": 33380 + }, + { + "epoch": 0.52083983278218, + "grad_norm": 3.800628662109375, + "learning_rate": 0.00016527838439300347, + "loss": 0.458, + "step": 33390 + }, + { + "epoch": 0.5209958195545018, + "grad_norm": 1.1253124475479126, + "learning_rate": 0.0001652679852748487, + "loss": 0.1238, + "step": 33400 + }, + { + "epoch": 0.5211518063268235, + "grad_norm": 0.46979108452796936, + "learning_rate": 0.0001652575861566939, + "loss": 0.0745, + "step": 33410 + }, + { + "epoch": 0.5213077930991452, + "grad_norm": 7.921492576599121, + "learning_rate": 0.00016524718703853913, + "loss": 0.4004, + "step": 33420 + }, + { + "epoch": 0.5214637798714669, + "grad_norm": 1.643127679824829, + "learning_rate": 0.00016523678792038435, + "loss": 0.2858, + "step": 33430 + }, + { + "epoch": 0.5216197666437886, + "grad_norm": 7.096631050109863, + "learning_rate": 0.0001652263888022296, + "loss": 0.4735, + "step": 33440 + }, + { + "epoch": 0.5217757534161103, + "grad_norm": 0.7280219197273254, + "learning_rate": 0.0001652159896840748, + "loss": 0.1841, + "step": 33450 + }, + { + "epoch": 0.521931740188432, + "grad_norm": 0.9415522813796997, + "learning_rate": 0.00016520559056592003, + "loss": 0.2169, + "step": 33460 + }, + { + "epoch": 0.5220877269607537, + "grad_norm": 2.264862060546875, + "learning_rate": 0.00016519519144776523, + "loss": 0.1902, + "step": 33470 + }, + { + "epoch": 0.5222437137330754, + "grad_norm": 2.4566822052001953, + "learning_rate": 0.00016518479232961047, + "loss": 0.3329, + "step": 33480 + }, + { + "epoch": 0.5223997005053972, + "grad_norm": 0.7330378890037537, + "learning_rate": 0.00016517439321145567, + "loss": 0.2341, + "step": 33490 + }, + { + "epoch": 0.5225556872777188, + "grad_norm": 1.6409542560577393, + "learning_rate": 0.0001651639940933009, + "loss": 0.3077, + "step": 33500 + }, + { + "epoch": 0.5227116740500406, + "grad_norm": 1.0647432804107666, + "learning_rate": 0.0001651535949751461, + "loss": 0.1935, + "step": 33510 + }, + { + "epoch": 0.5228676608223622, + "grad_norm": 0.507420539855957, + "learning_rate": 0.00016514319585699135, + "loss": 0.1314, + "step": 33520 + }, + { + "epoch": 0.523023647594684, + "grad_norm": 1.0064164400100708, + "learning_rate": 0.00016513279673883654, + "loss": 0.344, + "step": 33530 + }, + { + "epoch": 0.5231796343670057, + "grad_norm": 1.1936209201812744, + "learning_rate": 0.0001651223976206818, + "loss": 0.1948, + "step": 33540 + }, + { + "epoch": 0.5233356211393274, + "grad_norm": 0.04992926865816116, + "learning_rate": 0.00016511199850252698, + "loss": 0.1363, + "step": 33550 + }, + { + "epoch": 0.5234916079116491, + "grad_norm": 1.8256275653839111, + "learning_rate": 0.00016510159938437223, + "loss": 0.443, + "step": 33560 + }, + { + "epoch": 0.5236475946839708, + "grad_norm": 4.914908409118652, + "learning_rate": 0.00016509120026621742, + "loss": 0.4654, + "step": 33570 + }, + { + "epoch": 0.5238035814562925, + "grad_norm": 1.1625486612319946, + "learning_rate": 0.00016508080114806267, + "loss": 0.1974, + "step": 33580 + }, + { + "epoch": 0.5239595682286142, + "grad_norm": 6.6192522048950195, + "learning_rate": 0.00016507040202990786, + "loss": 0.2131, + "step": 33590 + }, + { + "epoch": 0.5241155550009359, + "grad_norm": 0.7369065880775452, + "learning_rate": 0.0001650600029117531, + "loss": 0.2371, + "step": 33600 + }, + { + "epoch": 0.5242715417732576, + "grad_norm": 1.5238152742385864, + "learning_rate": 0.0001650496037935983, + "loss": 0.1557, + "step": 33610 + }, + { + "epoch": 0.5244275285455793, + "grad_norm": 1.0418007373809814, + "learning_rate": 0.00016503920467544355, + "loss": 0.1878, + "step": 33620 + }, + { + "epoch": 0.524583515317901, + "grad_norm": 0.790117084980011, + "learning_rate": 0.00016502880555728874, + "loss": 0.2195, + "step": 33630 + }, + { + "epoch": 0.5247395020902228, + "grad_norm": 1.6712257862091064, + "learning_rate": 0.000165018406439134, + "loss": 0.1602, + "step": 33640 + }, + { + "epoch": 0.5248954888625444, + "grad_norm": 0.19236230850219727, + "learning_rate": 0.00016500800732097918, + "loss": 0.2526, + "step": 33650 + }, + { + "epoch": 0.5250514756348662, + "grad_norm": 1.3519701957702637, + "learning_rate": 0.00016499760820282443, + "loss": 0.4686, + "step": 33660 + }, + { + "epoch": 0.5252074624071879, + "grad_norm": 1.694342851638794, + "learning_rate": 0.00016498720908466962, + "loss": 0.1859, + "step": 33670 + }, + { + "epoch": 0.5253634491795096, + "grad_norm": 5.225239276885986, + "learning_rate": 0.00016497680996651487, + "loss": 0.2944, + "step": 33680 + }, + { + "epoch": 0.5255194359518313, + "grad_norm": 2.0208842754364014, + "learning_rate": 0.00016496641084836006, + "loss": 0.2421, + "step": 33690 + }, + { + "epoch": 0.525675422724153, + "grad_norm": 0.7954996228218079, + "learning_rate": 0.0001649560117302053, + "loss": 0.2853, + "step": 33700 + }, + { + "epoch": 0.5258314094964747, + "grad_norm": 2.296086072921753, + "learning_rate": 0.0001649456126120505, + "loss": 0.1333, + "step": 33710 + }, + { + "epoch": 0.5259873962687964, + "grad_norm": 1.1779128313064575, + "learning_rate": 0.00016493521349389574, + "loss": 0.1066, + "step": 33720 + }, + { + "epoch": 0.5261433830411181, + "grad_norm": 0.1756065934896469, + "learning_rate": 0.00016492481437574094, + "loss": 0.1352, + "step": 33730 + }, + { + "epoch": 0.5262993698134398, + "grad_norm": 0.13100725412368774, + "learning_rate": 0.00016491441525758618, + "loss": 0.2399, + "step": 33740 + }, + { + "epoch": 0.5264553565857615, + "grad_norm": 5.532008171081543, + "learning_rate": 0.00016490401613943138, + "loss": 0.2896, + "step": 33750 + }, + { + "epoch": 0.5266113433580832, + "grad_norm": 1.319886565208435, + "learning_rate": 0.00016489361702127662, + "loss": 0.3275, + "step": 33760 + }, + { + "epoch": 0.5267673301304049, + "grad_norm": 1.5550974607467651, + "learning_rate": 0.00016488321790312182, + "loss": 0.2677, + "step": 33770 + }, + { + "epoch": 0.5269233169027266, + "grad_norm": 1.8936737775802612, + "learning_rate": 0.00016487281878496706, + "loss": 0.1955, + "step": 33780 + }, + { + "epoch": 0.5270793036750484, + "grad_norm": 0.3653401732444763, + "learning_rate": 0.00016486241966681226, + "loss": 0.0723, + "step": 33790 + }, + { + "epoch": 0.52723529044737, + "grad_norm": 2.861341714859009, + "learning_rate": 0.00016485202054865747, + "loss": 0.2412, + "step": 33800 + }, + { + "epoch": 0.5273912772196918, + "grad_norm": 1.5291428565979004, + "learning_rate": 0.0001648416214305027, + "loss": 0.0871, + "step": 33810 + }, + { + "epoch": 0.5275472639920135, + "grad_norm": 1.0372581481933594, + "learning_rate": 0.00016483122231234791, + "loss": 0.4705, + "step": 33820 + }, + { + "epoch": 0.5277032507643352, + "grad_norm": 1.1943141222000122, + "learning_rate": 0.00016482082319419313, + "loss": 0.2848, + "step": 33830 + }, + { + "epoch": 0.5278592375366569, + "grad_norm": 1.9008225202560425, + "learning_rate": 0.00016481042407603835, + "loss": 0.139, + "step": 33840 + }, + { + "epoch": 0.5280152243089786, + "grad_norm": 2.132089138031006, + "learning_rate": 0.00016480002495788357, + "loss": 0.2119, + "step": 33850 + }, + { + "epoch": 0.5281712110813003, + "grad_norm": 0.24524426460266113, + "learning_rate": 0.0001647896258397288, + "loss": 0.1643, + "step": 33860 + }, + { + "epoch": 0.528327197853622, + "grad_norm": 1.6469637155532837, + "learning_rate": 0.000164779226721574, + "loss": 0.363, + "step": 33870 + }, + { + "epoch": 0.5284831846259437, + "grad_norm": 0.8767328858375549, + "learning_rate": 0.00016476882760341923, + "loss": 0.2632, + "step": 33880 + }, + { + "epoch": 0.5286391713982654, + "grad_norm": 0.06347586214542389, + "learning_rate": 0.00016475842848526445, + "loss": 0.3204, + "step": 33890 + }, + { + "epoch": 0.5287951581705871, + "grad_norm": 0.09782540798187256, + "learning_rate": 0.00016474802936710967, + "loss": 0.2211, + "step": 33900 + }, + { + "epoch": 0.5289511449429088, + "grad_norm": 3.2998859882354736, + "learning_rate": 0.0001647376302489549, + "loss": 0.265, + "step": 33910 + }, + { + "epoch": 0.5291071317152305, + "grad_norm": 0.43594226241111755, + "learning_rate": 0.0001647272311308001, + "loss": 0.194, + "step": 33920 + }, + { + "epoch": 0.5292631184875523, + "grad_norm": 1.5166605710983276, + "learning_rate": 0.00016471683201264533, + "loss": 0.2675, + "step": 33930 + }, + { + "epoch": 0.529419105259874, + "grad_norm": 0.6056640148162842, + "learning_rate": 0.00016470643289449055, + "loss": 0.0576, + "step": 33940 + }, + { + "epoch": 0.5295750920321957, + "grad_norm": 0.25410348176956177, + "learning_rate": 0.00016469603377633577, + "loss": 0.1124, + "step": 33950 + }, + { + "epoch": 0.5297310788045174, + "grad_norm": 1.770642876625061, + "learning_rate": 0.000164685634658181, + "loss": 0.3295, + "step": 33960 + }, + { + "epoch": 0.5298870655768391, + "grad_norm": 0.0607205331325531, + "learning_rate": 0.0001646752355400262, + "loss": 0.2369, + "step": 33970 + }, + { + "epoch": 0.5300430523491608, + "grad_norm": 0.5557095408439636, + "learning_rate": 0.00016466483642187143, + "loss": 0.1681, + "step": 33980 + }, + { + "epoch": 0.5301990391214825, + "grad_norm": 0.5192957520484924, + "learning_rate": 0.00016465443730371665, + "loss": 0.2649, + "step": 33990 + }, + { + "epoch": 0.5303550258938042, + "grad_norm": 0.04804835096001625, + "learning_rate": 0.00016464403818556187, + "loss": 0.2639, + "step": 34000 + }, + { + "epoch": 0.5305110126661259, + "grad_norm": 0.02673129364848137, + "learning_rate": 0.0001646336390674071, + "loss": 0.169, + "step": 34010 + }, + { + "epoch": 0.5306669994384476, + "grad_norm": 1.0084244012832642, + "learning_rate": 0.0001646232399492523, + "loss": 0.353, + "step": 34020 + }, + { + "epoch": 0.5308229862107693, + "grad_norm": 2.2202091217041016, + "learning_rate": 0.00016461284083109753, + "loss": 0.3206, + "step": 34030 + }, + { + "epoch": 0.530978972983091, + "grad_norm": 0.5573744773864746, + "learning_rate": 0.00016460244171294275, + "loss": 0.231, + "step": 34040 + }, + { + "epoch": 0.5311349597554127, + "grad_norm": 0.39700084924697876, + "learning_rate": 0.00016459204259478797, + "loss": 0.1412, + "step": 34050 + }, + { + "epoch": 0.5312909465277345, + "grad_norm": 2.582963228225708, + "learning_rate": 0.00016458164347663319, + "loss": 0.3334, + "step": 34060 + }, + { + "epoch": 0.5314469333000561, + "grad_norm": 0.22781169414520264, + "learning_rate": 0.0001645712443584784, + "loss": 0.2767, + "step": 34070 + }, + { + "epoch": 0.5316029200723779, + "grad_norm": 1.2051042318344116, + "learning_rate": 0.00016456084524032362, + "loss": 0.2045, + "step": 34080 + }, + { + "epoch": 0.5317589068446996, + "grad_norm": 0.42760100960731506, + "learning_rate": 0.00016455044612216884, + "loss": 0.1151, + "step": 34090 + }, + { + "epoch": 0.5319148936170213, + "grad_norm": 0.14440476894378662, + "learning_rate": 0.00016454004700401406, + "loss": 0.2386, + "step": 34100 + }, + { + "epoch": 0.532070880389343, + "grad_norm": 2.2777981758117676, + "learning_rate": 0.00016452964788585928, + "loss": 0.2009, + "step": 34110 + }, + { + "epoch": 0.5322268671616647, + "grad_norm": 0.9206979274749756, + "learning_rate": 0.0001645192487677045, + "loss": 0.2745, + "step": 34120 + }, + { + "epoch": 0.5323828539339864, + "grad_norm": 1.6947574615478516, + "learning_rate": 0.00016450884964954972, + "loss": 0.2584, + "step": 34130 + }, + { + "epoch": 0.5325388407063081, + "grad_norm": 0.401444673538208, + "learning_rate": 0.00016449845053139494, + "loss": 0.2218, + "step": 34140 + }, + { + "epoch": 0.5326948274786298, + "grad_norm": 0.08261553198099136, + "learning_rate": 0.00016448805141324016, + "loss": 0.2775, + "step": 34150 + }, + { + "epoch": 0.5328508142509515, + "grad_norm": 0.1017974391579628, + "learning_rate": 0.00016447765229508538, + "loss": 0.2095, + "step": 34160 + }, + { + "epoch": 0.5330068010232732, + "grad_norm": 1.3759571313858032, + "learning_rate": 0.0001644672531769306, + "loss": 0.2643, + "step": 34170 + }, + { + "epoch": 0.5331627877955949, + "grad_norm": 1.2654389142990112, + "learning_rate": 0.00016445685405877582, + "loss": 0.2949, + "step": 34180 + }, + { + "epoch": 0.5333187745679167, + "grad_norm": 1.5481843948364258, + "learning_rate": 0.00016444645494062104, + "loss": 0.1264, + "step": 34190 + }, + { + "epoch": 0.5334747613402383, + "grad_norm": 1.8094528913497925, + "learning_rate": 0.00016443605582246626, + "loss": 0.2727, + "step": 34200 + }, + { + "epoch": 0.5336307481125601, + "grad_norm": 2.224538564682007, + "learning_rate": 0.00016442565670431148, + "loss": 0.3096, + "step": 34210 + }, + { + "epoch": 0.5337867348848817, + "grad_norm": 0.6375226974487305, + "learning_rate": 0.0001644152575861567, + "loss": 0.2251, + "step": 34220 + }, + { + "epoch": 0.5339427216572035, + "grad_norm": 3.727106809616089, + "learning_rate": 0.00016440485846800192, + "loss": 0.4374, + "step": 34230 + }, + { + "epoch": 0.5340987084295252, + "grad_norm": 0.13345426321029663, + "learning_rate": 0.00016439445934984714, + "loss": 0.2011, + "step": 34240 + }, + { + "epoch": 0.5342546952018469, + "grad_norm": 2.1658668518066406, + "learning_rate": 0.00016438406023169236, + "loss": 0.2457, + "step": 34250 + }, + { + "epoch": 0.5344106819741686, + "grad_norm": 19.238407135009766, + "learning_rate": 0.00016437366111353758, + "loss": 0.2756, + "step": 34260 + }, + { + "epoch": 0.5345666687464903, + "grad_norm": 1.0292778015136719, + "learning_rate": 0.0001643632619953828, + "loss": 0.1646, + "step": 34270 + }, + { + "epoch": 0.534722655518812, + "grad_norm": 0.9372987747192383, + "learning_rate": 0.00016435286287722802, + "loss": 0.2762, + "step": 34280 + }, + { + "epoch": 0.5348786422911337, + "grad_norm": 0.3918002247810364, + "learning_rate": 0.00016434246375907324, + "loss": 0.1406, + "step": 34290 + }, + { + "epoch": 0.5350346290634554, + "grad_norm": 1.3518732786178589, + "learning_rate": 0.00016433206464091846, + "loss": 0.2992, + "step": 34300 + }, + { + "epoch": 0.5351906158357771, + "grad_norm": 0.73117595911026, + "learning_rate": 0.00016432166552276368, + "loss": 0.1172, + "step": 34310 + }, + { + "epoch": 0.5353466026080989, + "grad_norm": 0.06655958294868469, + "learning_rate": 0.0001643112664046089, + "loss": 0.1332, + "step": 34320 + }, + { + "epoch": 0.5355025893804205, + "grad_norm": 3.2587168216705322, + "learning_rate": 0.00016430086728645412, + "loss": 0.2052, + "step": 34330 + }, + { + "epoch": 0.5356585761527423, + "grad_norm": 2.6668431758880615, + "learning_rate": 0.00016429046816829934, + "loss": 0.1295, + "step": 34340 + }, + { + "epoch": 0.5358145629250639, + "grad_norm": 1.316080927848816, + "learning_rate": 0.00016428006905014456, + "loss": 0.2072, + "step": 34350 + }, + { + "epoch": 0.5359705496973857, + "grad_norm": 5.144528388977051, + "learning_rate": 0.00016426966993198977, + "loss": 0.1934, + "step": 34360 + }, + { + "epoch": 0.5361265364697073, + "grad_norm": 1.3529599905014038, + "learning_rate": 0.000164259270813835, + "loss": 0.2077, + "step": 34370 + }, + { + "epoch": 0.5362825232420291, + "grad_norm": 0.9132925868034363, + "learning_rate": 0.00016424887169568021, + "loss": 0.2404, + "step": 34380 + }, + { + "epoch": 0.5364385100143508, + "grad_norm": 1.520033597946167, + "learning_rate": 0.00016423847257752543, + "loss": 0.1765, + "step": 34390 + }, + { + "epoch": 0.5365944967866725, + "grad_norm": 2.343975305557251, + "learning_rate": 0.00016422807345937065, + "loss": 0.1959, + "step": 34400 + }, + { + "epoch": 0.5367504835589942, + "grad_norm": 0.5855118632316589, + "learning_rate": 0.00016421767434121587, + "loss": 0.0848, + "step": 34410 + }, + { + "epoch": 0.5369064703313159, + "grad_norm": 1.9108648300170898, + "learning_rate": 0.0001642072752230611, + "loss": 0.2406, + "step": 34420 + }, + { + "epoch": 0.5370624571036376, + "grad_norm": 1.0323792695999146, + "learning_rate": 0.0001641968761049063, + "loss": 0.2298, + "step": 34430 + }, + { + "epoch": 0.5372184438759593, + "grad_norm": 0.17435620725154877, + "learning_rate": 0.00016418647698675153, + "loss": 0.1396, + "step": 34440 + }, + { + "epoch": 0.537374430648281, + "grad_norm": 0.3526577353477478, + "learning_rate": 0.00016417607786859675, + "loss": 0.172, + "step": 34450 + }, + { + "epoch": 0.5375304174206027, + "grad_norm": 2.73685884475708, + "learning_rate": 0.00016416567875044197, + "loss": 0.1167, + "step": 34460 + }, + { + "epoch": 0.5376864041929245, + "grad_norm": 2.765693426132202, + "learning_rate": 0.0001641552796322872, + "loss": 0.1356, + "step": 34470 + }, + { + "epoch": 0.5378423909652461, + "grad_norm": 0.47628021240234375, + "learning_rate": 0.0001641448805141324, + "loss": 0.3104, + "step": 34480 + }, + { + "epoch": 0.5379983777375679, + "grad_norm": 1.1307590007781982, + "learning_rate": 0.00016413448139597763, + "loss": 0.3247, + "step": 34490 + }, + { + "epoch": 0.5381543645098895, + "grad_norm": 0.07136381417512894, + "learning_rate": 0.00016412408227782285, + "loss": 0.319, + "step": 34500 + }, + { + "epoch": 0.5383103512822113, + "grad_norm": 1.7850221395492554, + "learning_rate": 0.00016411368315966807, + "loss": 0.2515, + "step": 34510 + }, + { + "epoch": 0.5384663380545329, + "grad_norm": 1.3870742321014404, + "learning_rate": 0.0001641032840415133, + "loss": 0.256, + "step": 34520 + }, + { + "epoch": 0.5386223248268547, + "grad_norm": 1.1329221725463867, + "learning_rate": 0.0001640928849233585, + "loss": 0.2255, + "step": 34530 + }, + { + "epoch": 0.5387783115991764, + "grad_norm": 0.23236137628555298, + "learning_rate": 0.00016408248580520373, + "loss": 0.4053, + "step": 34540 + }, + { + "epoch": 0.5389342983714981, + "grad_norm": 3.350924253463745, + "learning_rate": 0.00016407208668704895, + "loss": 0.2816, + "step": 34550 + }, + { + "epoch": 0.5390902851438198, + "grad_norm": 1.7459170818328857, + "learning_rate": 0.00016406168756889417, + "loss": 0.0834, + "step": 34560 + }, + { + "epoch": 0.5392462719161415, + "grad_norm": 0.89351487159729, + "learning_rate": 0.0001640512884507394, + "loss": 0.1146, + "step": 34570 + }, + { + "epoch": 0.5394022586884633, + "grad_norm": 0.49410831928253174, + "learning_rate": 0.0001640408893325846, + "loss": 0.2319, + "step": 34580 + }, + { + "epoch": 0.5395582454607849, + "grad_norm": 0.009971237741410732, + "learning_rate": 0.00016403049021442983, + "loss": 0.1674, + "step": 34590 + }, + { + "epoch": 0.5397142322331067, + "grad_norm": 1.420511245727539, + "learning_rate": 0.00016402009109627505, + "loss": 0.1469, + "step": 34600 + }, + { + "epoch": 0.5398702190054283, + "grad_norm": 2.5686769485473633, + "learning_rate": 0.00016400969197812027, + "loss": 0.2607, + "step": 34610 + }, + { + "epoch": 0.5400262057777501, + "grad_norm": 2.5808913707733154, + "learning_rate": 0.00016399929285996549, + "loss": 0.4412, + "step": 34620 + }, + { + "epoch": 0.5401821925500717, + "grad_norm": 0.2778591811656952, + "learning_rate": 0.0001639888937418107, + "loss": 0.1563, + "step": 34630 + }, + { + "epoch": 0.5403381793223935, + "grad_norm": 1.6448099613189697, + "learning_rate": 0.00016397849462365592, + "loss": 0.2484, + "step": 34640 + }, + { + "epoch": 0.5404941660947151, + "grad_norm": 0.003213417250663042, + "learning_rate": 0.00016396809550550114, + "loss": 0.1582, + "step": 34650 + }, + { + "epoch": 0.5406501528670369, + "grad_norm": 2.3433990478515625, + "learning_rate": 0.00016395769638734636, + "loss": 0.2774, + "step": 34660 + }, + { + "epoch": 0.5408061396393585, + "grad_norm": 0.09544426947832108, + "learning_rate": 0.00016394729726919158, + "loss": 0.1242, + "step": 34670 + }, + { + "epoch": 0.5409621264116803, + "grad_norm": 1.8934235572814941, + "learning_rate": 0.0001639368981510368, + "loss": 0.1636, + "step": 34680 + }, + { + "epoch": 0.541118113184002, + "grad_norm": 2.8528554439544678, + "learning_rate": 0.00016392649903288202, + "loss": 0.2524, + "step": 34690 + }, + { + "epoch": 0.5412740999563237, + "grad_norm": 2.837836980819702, + "learning_rate": 0.00016391609991472722, + "loss": 0.1808, + "step": 34700 + }, + { + "epoch": 0.5414300867286455, + "grad_norm": 4.349740028381348, + "learning_rate": 0.00016390570079657246, + "loss": 0.12, + "step": 34710 + }, + { + "epoch": 0.5415860735009671, + "grad_norm": 0.6011945605278015, + "learning_rate": 0.00016389530167841765, + "loss": 0.2865, + "step": 34720 + }, + { + "epoch": 0.5417420602732889, + "grad_norm": 1.2079488039016724, + "learning_rate": 0.0001638849025602629, + "loss": 0.125, + "step": 34730 + }, + { + "epoch": 0.5418980470456105, + "grad_norm": 1.9549164772033691, + "learning_rate": 0.0001638745034421081, + "loss": 0.3285, + "step": 34740 + }, + { + "epoch": 0.5420540338179323, + "grad_norm": 0.6969407200813293, + "learning_rate": 0.00016386410432395334, + "loss": 0.3267, + "step": 34750 + }, + { + "epoch": 0.5422100205902539, + "grad_norm": 1.0561965703964233, + "learning_rate": 0.00016385370520579853, + "loss": 0.1463, + "step": 34760 + }, + { + "epoch": 0.5423660073625757, + "grad_norm": 1.8208931684494019, + "learning_rate": 0.00016384330608764378, + "loss": 0.1713, + "step": 34770 + }, + { + "epoch": 0.5425219941348973, + "grad_norm": 0.6349910497665405, + "learning_rate": 0.000163832906969489, + "loss": 0.2244, + "step": 34780 + }, + { + "epoch": 0.5426779809072191, + "grad_norm": 1.1976515054702759, + "learning_rate": 0.00016382250785133422, + "loss": 0.3407, + "step": 34790 + }, + { + "epoch": 0.5428339676795407, + "grad_norm": 0.5970319509506226, + "learning_rate": 0.00016381210873317944, + "loss": 0.1419, + "step": 34800 + }, + { + "epoch": 0.5429899544518625, + "grad_norm": 0.14623159170150757, + "learning_rate": 0.00016380170961502466, + "loss": 0.14, + "step": 34810 + }, + { + "epoch": 0.5431459412241841, + "grad_norm": 1.4332351684570312, + "learning_rate": 0.00016379131049686988, + "loss": 0.1093, + "step": 34820 + }, + { + "epoch": 0.5433019279965059, + "grad_norm": 1.1443063020706177, + "learning_rate": 0.0001637809113787151, + "loss": 0.3676, + "step": 34830 + }, + { + "epoch": 0.5434579147688277, + "grad_norm": 1.3897461891174316, + "learning_rate": 0.00016377051226056032, + "loss": 0.3713, + "step": 34840 + }, + { + "epoch": 0.5436139015411493, + "grad_norm": 1.7773199081420898, + "learning_rate": 0.00016376011314240554, + "loss": 0.2249, + "step": 34850 + }, + { + "epoch": 0.5437698883134711, + "grad_norm": 0.8524389266967773, + "learning_rate": 0.00016374971402425076, + "loss": 0.2158, + "step": 34860 + }, + { + "epoch": 0.5439258750857927, + "grad_norm": 0.5572389960289001, + "learning_rate": 0.00016373931490609598, + "loss": 0.215, + "step": 34870 + }, + { + "epoch": 0.5440818618581145, + "grad_norm": 3.0238709449768066, + "learning_rate": 0.0001637289157879412, + "loss": 0.3765, + "step": 34880 + }, + { + "epoch": 0.5442378486304361, + "grad_norm": 1.1655421257019043, + "learning_rate": 0.00016371851666978642, + "loss": 0.2491, + "step": 34890 + }, + { + "epoch": 0.5443938354027579, + "grad_norm": 0.49492090940475464, + "learning_rate": 0.00016370811755163164, + "loss": 0.1287, + "step": 34900 + }, + { + "epoch": 0.5445498221750795, + "grad_norm": 0.2732921540737152, + "learning_rate": 0.00016369771843347685, + "loss": 0.2154, + "step": 34910 + }, + { + "epoch": 0.5447058089474013, + "grad_norm": 2.5807769298553467, + "learning_rate": 0.00016368731931532207, + "loss": 0.2032, + "step": 34920 + }, + { + "epoch": 0.5448617957197229, + "grad_norm": 0.5044315457344055, + "learning_rate": 0.0001636769201971673, + "loss": 0.2595, + "step": 34930 + }, + { + "epoch": 0.5450177824920447, + "grad_norm": 2.1248972415924072, + "learning_rate": 0.00016366652107901251, + "loss": 0.5651, + "step": 34940 + }, + { + "epoch": 0.5451737692643663, + "grad_norm": 0.8391468524932861, + "learning_rate": 0.00016365612196085773, + "loss": 0.3028, + "step": 34950 + }, + { + "epoch": 0.5453297560366881, + "grad_norm": 0.36081477999687195, + "learning_rate": 0.00016364572284270295, + "loss": 0.1073, + "step": 34960 + }, + { + "epoch": 0.5454857428090097, + "grad_norm": 3.175804376602173, + "learning_rate": 0.00016363532372454817, + "loss": 0.1889, + "step": 34970 + }, + { + "epoch": 0.5456417295813315, + "grad_norm": 0.20647937059402466, + "learning_rate": 0.0001636249246063934, + "loss": 0.2416, + "step": 34980 + }, + { + "epoch": 0.5457977163536533, + "grad_norm": 2.3525943756103516, + "learning_rate": 0.0001636145254882386, + "loss": 0.3179, + "step": 34990 + }, + { + "epoch": 0.5459537031259749, + "grad_norm": 0.2641216814517975, + "learning_rate": 0.00016360412637008383, + "loss": 0.2157, + "step": 35000 + }, + { + "epoch": 0.5461096898982967, + "grad_norm": 0.09906073659658432, + "learning_rate": 0.00016359372725192905, + "loss": 0.2576, + "step": 35010 + }, + { + "epoch": 0.5462656766706183, + "grad_norm": 0.13971920311450958, + "learning_rate": 0.00016358332813377427, + "loss": 0.1298, + "step": 35020 + }, + { + "epoch": 0.5464216634429401, + "grad_norm": 1.4535638093948364, + "learning_rate": 0.0001635729290156195, + "loss": 0.2357, + "step": 35030 + }, + { + "epoch": 0.5465776502152617, + "grad_norm": 0.9107828736305237, + "learning_rate": 0.0001635625298974647, + "loss": 0.3049, + "step": 35040 + }, + { + "epoch": 0.5467336369875835, + "grad_norm": 0.0377386212348938, + "learning_rate": 0.00016355213077930993, + "loss": 0.2066, + "step": 35050 + }, + { + "epoch": 0.5468896237599051, + "grad_norm": 2.7800869941711426, + "learning_rate": 0.00016354173166115515, + "loss": 0.2974, + "step": 35060 + }, + { + "epoch": 0.5470456105322269, + "grad_norm": 1.6247998476028442, + "learning_rate": 0.00016353133254300037, + "loss": 0.3822, + "step": 35070 + }, + { + "epoch": 0.5472015973045485, + "grad_norm": 1.6479015350341797, + "learning_rate": 0.0001635209334248456, + "loss": 0.3401, + "step": 35080 + }, + { + "epoch": 0.5473575840768703, + "grad_norm": 0.1879737675189972, + "learning_rate": 0.0001635105343066908, + "loss": 0.1549, + "step": 35090 + }, + { + "epoch": 0.5475135708491919, + "grad_norm": 0.39355361461639404, + "learning_rate": 0.00016350013518853603, + "loss": 0.2032, + "step": 35100 + }, + { + "epoch": 0.5476695576215137, + "grad_norm": 2.9976983070373535, + "learning_rate": 0.00016348973607038125, + "loss": 0.3867, + "step": 35110 + }, + { + "epoch": 0.5478255443938354, + "grad_norm": 0.8242707252502441, + "learning_rate": 0.00016347933695222647, + "loss": 0.3601, + "step": 35120 + }, + { + "epoch": 0.5479815311661571, + "grad_norm": 0.874955415725708, + "learning_rate": 0.0001634689378340717, + "loss": 0.2709, + "step": 35130 + }, + { + "epoch": 0.5481375179384789, + "grad_norm": 0.05980971083045006, + "learning_rate": 0.0001634585387159169, + "loss": 0.2646, + "step": 35140 + }, + { + "epoch": 0.5482935047108005, + "grad_norm": 1.8437328338623047, + "learning_rate": 0.0001634481395977621, + "loss": 0.1965, + "step": 35150 + }, + { + "epoch": 0.5484494914831223, + "grad_norm": 0.4142405688762665, + "learning_rate": 0.00016343774047960735, + "loss": 0.1756, + "step": 35160 + }, + { + "epoch": 0.5486054782554439, + "grad_norm": 4.836324214935303, + "learning_rate": 0.00016342734136145254, + "loss": 0.2534, + "step": 35170 + }, + { + "epoch": 0.5487614650277657, + "grad_norm": 1.4933065176010132, + "learning_rate": 0.00016341694224329779, + "loss": 0.2355, + "step": 35180 + }, + { + "epoch": 0.5489174518000873, + "grad_norm": 1.9468894004821777, + "learning_rate": 0.00016340654312514298, + "loss": 0.3113, + "step": 35190 + }, + { + "epoch": 0.5490734385724091, + "grad_norm": 1.012710452079773, + "learning_rate": 0.00016339614400698822, + "loss": 0.2635, + "step": 35200 + }, + { + "epoch": 0.5492294253447307, + "grad_norm": 1.5551140308380127, + "learning_rate": 0.00016338574488883342, + "loss": 0.2951, + "step": 35210 + }, + { + "epoch": 0.5493854121170525, + "grad_norm": 1.8613696098327637, + "learning_rate": 0.00016337534577067866, + "loss": 0.2832, + "step": 35220 + }, + { + "epoch": 0.5495413988893741, + "grad_norm": 1.6748839616775513, + "learning_rate": 0.00016336494665252386, + "loss": 0.2266, + "step": 35230 + }, + { + "epoch": 0.5496973856616959, + "grad_norm": 0.4094032347202301, + "learning_rate": 0.0001633545475343691, + "loss": 0.3497, + "step": 35240 + }, + { + "epoch": 0.5498533724340176, + "grad_norm": 1.6414631605148315, + "learning_rate": 0.0001633441484162143, + "loss": 0.1763, + "step": 35250 + }, + { + "epoch": 0.5500093592063393, + "grad_norm": 1.6380645036697388, + "learning_rate": 0.00016333374929805954, + "loss": 0.1355, + "step": 35260 + }, + { + "epoch": 0.550165345978661, + "grad_norm": 0.7325630187988281, + "learning_rate": 0.00016332335017990473, + "loss": 0.3105, + "step": 35270 + }, + { + "epoch": 0.5503213327509827, + "grad_norm": 1.6015644073486328, + "learning_rate": 0.00016331295106174998, + "loss": 0.2311, + "step": 35280 + }, + { + "epoch": 0.5504773195233045, + "grad_norm": 0.9224210381507874, + "learning_rate": 0.00016330255194359517, + "loss": 0.2746, + "step": 35290 + }, + { + "epoch": 0.5506333062956261, + "grad_norm": 1.30025315284729, + "learning_rate": 0.00016329215282544042, + "loss": 0.2066, + "step": 35300 + }, + { + "epoch": 0.5507892930679479, + "grad_norm": 0.481125146150589, + "learning_rate": 0.0001632817537072856, + "loss": 0.1778, + "step": 35310 + }, + { + "epoch": 0.5509452798402695, + "grad_norm": 0.2709486782550812, + "learning_rate": 0.00016327135458913086, + "loss": 0.1131, + "step": 35320 + }, + { + "epoch": 0.5511012666125913, + "grad_norm": 0.6248563528060913, + "learning_rate": 0.00016326095547097605, + "loss": 0.4356, + "step": 35330 + }, + { + "epoch": 0.5512572533849129, + "grad_norm": 3.0947647094726562, + "learning_rate": 0.0001632505563528213, + "loss": 0.3072, + "step": 35340 + }, + { + "epoch": 0.5514132401572347, + "grad_norm": 1.009535312652588, + "learning_rate": 0.0001632401572346665, + "loss": 0.1808, + "step": 35350 + }, + { + "epoch": 0.5515692269295563, + "grad_norm": 0.042604975402355194, + "learning_rate": 0.00016322975811651174, + "loss": 0.1198, + "step": 35360 + }, + { + "epoch": 0.5517252137018781, + "grad_norm": 1.8652396202087402, + "learning_rate": 0.00016321935899835693, + "loss": 0.2176, + "step": 35370 + }, + { + "epoch": 0.5518812004741998, + "grad_norm": 4.6097187995910645, + "learning_rate": 0.00016320895988020218, + "loss": 0.2132, + "step": 35380 + }, + { + "epoch": 0.5520371872465215, + "grad_norm": 1.3615522384643555, + "learning_rate": 0.00016319856076204737, + "loss": 0.2751, + "step": 35390 + }, + { + "epoch": 0.5521931740188432, + "grad_norm": 0.8844773769378662, + "learning_rate": 0.00016318816164389262, + "loss": 0.22, + "step": 35400 + }, + { + "epoch": 0.5523491607911649, + "grad_norm": 0.5413331985473633, + "learning_rate": 0.0001631777625257378, + "loss": 0.2159, + "step": 35410 + }, + { + "epoch": 0.5525051475634866, + "grad_norm": 1.5456678867340088, + "learning_rate": 0.00016316736340758306, + "loss": 0.2244, + "step": 35420 + }, + { + "epoch": 0.5526611343358083, + "grad_norm": 2.1405861377716064, + "learning_rate": 0.00016315696428942825, + "loss": 0.446, + "step": 35430 + }, + { + "epoch": 0.5528171211081301, + "grad_norm": 1.4269858598709106, + "learning_rate": 0.0001631465651712735, + "loss": 0.3599, + "step": 35440 + }, + { + "epoch": 0.5529731078804517, + "grad_norm": 1.8682516813278198, + "learning_rate": 0.0001631361660531187, + "loss": 0.115, + "step": 35450 + }, + { + "epoch": 0.5531290946527735, + "grad_norm": 1.0175774097442627, + "learning_rate": 0.00016312576693496394, + "loss": 0.4003, + "step": 35460 + }, + { + "epoch": 0.5532850814250951, + "grad_norm": 2.4369170665740967, + "learning_rate": 0.00016311536781680913, + "loss": 0.2489, + "step": 35470 + }, + { + "epoch": 0.5534410681974169, + "grad_norm": 1.886022686958313, + "learning_rate": 0.00016310496869865437, + "loss": 0.4201, + "step": 35480 + }, + { + "epoch": 0.5535970549697385, + "grad_norm": 1.5811959505081177, + "learning_rate": 0.00016309456958049957, + "loss": 0.4286, + "step": 35490 + }, + { + "epoch": 0.5537530417420603, + "grad_norm": 1.971110224723816, + "learning_rate": 0.00016308417046234481, + "loss": 0.3599, + "step": 35500 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.161800511488e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-35500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3f9d8272fc461903512f795d348e9cfe78a53291 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:197f97c9af833d6adae363a2e83defd08a41dfc44a55cdcb57331d6e17370e58 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..e3a373c6efa5fe47ea7b1bcfd3a0072321c4bf99 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d53937790a60832a48914aa002805b891fccf55d23a3fcf7202c393023b68cc +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..3f1fa7b7954c9d3bcf850e4b70ee390dde8952e7 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..9582b6f5653b4e690dd64b8b1629caec0bb6fa60 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..cad92a2419c6cee7784617a5a009286d04e1edb7 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/trainer_state.json @@ -0,0 +1,25234 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5615523803581456, + "eval_steps": 500, + "global_step": 36000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + }, + { + "epoch": 0.5071129968178698, + "grad_norm": 1.0075199604034424, + "learning_rate": 0.00016619350679062414, + "loss": 0.2569, + "step": 32510 + }, + { + "epoch": 0.5072689835901916, + "grad_norm": 1.3489066362380981, + "learning_rate": 0.0001661831076724694, + "loss": 0.2565, + "step": 32520 + }, + { + "epoch": 0.5074249703625132, + "grad_norm": 0.2534058094024658, + "learning_rate": 0.00016617270855431458, + "loss": 0.2292, + "step": 32530 + }, + { + "epoch": 0.507580957134835, + "grad_norm": 0.31723281741142273, + "learning_rate": 0.00016616230943615983, + "loss": 0.1467, + "step": 32540 + }, + { + "epoch": 0.5077369439071566, + "grad_norm": 0.29659849405288696, + "learning_rate": 0.00016615191031800502, + "loss": 0.2768, + "step": 32550 + }, + { + "epoch": 0.5078929306794784, + "grad_norm": 1.5246435403823853, + "learning_rate": 0.00016614151119985027, + "loss": 0.216, + "step": 32560 + }, + { + "epoch": 0.5080489174518001, + "grad_norm": 0.6026854515075684, + "learning_rate": 0.00016613111208169546, + "loss": 0.2277, + "step": 32570 + }, + { + "epoch": 0.5082049042241218, + "grad_norm": 0.41994890570640564, + "learning_rate": 0.0001661207129635407, + "loss": 0.2433, + "step": 32580 + }, + { + "epoch": 0.5083608909964435, + "grad_norm": 2.084599733352661, + "learning_rate": 0.00016611031384538593, + "loss": 0.4081, + "step": 32590 + }, + { + "epoch": 0.5085168777687652, + "grad_norm": 0.2945229709148407, + "learning_rate": 0.00016609991472723115, + "loss": 0.1329, + "step": 32600 + }, + { + "epoch": 0.508672864541087, + "grad_norm": 1.4981423616409302, + "learning_rate": 0.00016608951560907636, + "loss": 0.2855, + "step": 32610 + }, + { + "epoch": 0.5088288513134086, + "grad_norm": 0.961737871170044, + "learning_rate": 0.00016607911649092158, + "loss": 0.1654, + "step": 32620 + }, + { + "epoch": 0.5089848380857304, + "grad_norm": 1.49734628200531, + "learning_rate": 0.0001660687173727668, + "loss": 0.2081, + "step": 32630 + }, + { + "epoch": 0.509140824858052, + "grad_norm": 2.538135051727295, + "learning_rate": 0.00016605831825461202, + "loss": 0.3023, + "step": 32640 + }, + { + "epoch": 0.5092968116303738, + "grad_norm": 0.2981272041797638, + "learning_rate": 0.00016604791913645724, + "loss": 0.214, + "step": 32650 + }, + { + "epoch": 0.5094527984026954, + "grad_norm": 1.8863650560379028, + "learning_rate": 0.00016603752001830246, + "loss": 0.3363, + "step": 32660 + }, + { + "epoch": 0.5096087851750172, + "grad_norm": 0.022720983251929283, + "learning_rate": 0.00016602712090014768, + "loss": 0.2463, + "step": 32670 + }, + { + "epoch": 0.5097647719473388, + "grad_norm": 1.154977798461914, + "learning_rate": 0.0001660167217819929, + "loss": 0.4484, + "step": 32680 + }, + { + "epoch": 0.5099207587196606, + "grad_norm": 0.9470499157905579, + "learning_rate": 0.00016600632266383812, + "loss": 0.1196, + "step": 32690 + }, + { + "epoch": 0.5100767454919822, + "grad_norm": 0.19955401122570038, + "learning_rate": 0.00016599592354568334, + "loss": 0.1762, + "step": 32700 + }, + { + "epoch": 0.510232732264304, + "grad_norm": 0.4098282754421234, + "learning_rate": 0.00016598552442752856, + "loss": 0.2739, + "step": 32710 + }, + { + "epoch": 0.5103887190366257, + "grad_norm": 0.21697258949279785, + "learning_rate": 0.00016597512530937378, + "loss": 0.2555, + "step": 32720 + }, + { + "epoch": 0.5105447058089474, + "grad_norm": 4.894479751586914, + "learning_rate": 0.000165964726191219, + "loss": 0.3137, + "step": 32730 + }, + { + "epoch": 0.5107006925812692, + "grad_norm": 0.8396711349487305, + "learning_rate": 0.00016595432707306422, + "loss": 0.3105, + "step": 32740 + }, + { + "epoch": 0.5108566793535908, + "grad_norm": 1.5494674444198608, + "learning_rate": 0.00016594392795490944, + "loss": 0.2271, + "step": 32750 + }, + { + "epoch": 0.5110126661259126, + "grad_norm": 1.3409888744354248, + "learning_rate": 0.00016593352883675466, + "loss": 0.1796, + "step": 32760 + }, + { + "epoch": 0.5111686528982342, + "grad_norm": 1.7590911388397217, + "learning_rate": 0.00016592312971859988, + "loss": 0.2185, + "step": 32770 + }, + { + "epoch": 0.511324639670556, + "grad_norm": 0.7350002527236938, + "learning_rate": 0.0001659127306004451, + "loss": 0.1911, + "step": 32780 + }, + { + "epoch": 0.5114806264428776, + "grad_norm": 0.7176411747932434, + "learning_rate": 0.00016590233148229032, + "loss": 0.6918, + "step": 32790 + }, + { + "epoch": 0.5116366132151994, + "grad_norm": 2.5005321502685547, + "learning_rate": 0.00016589193236413554, + "loss": 0.2329, + "step": 32800 + }, + { + "epoch": 0.511792599987521, + "grad_norm": 1.3916250467300415, + "learning_rate": 0.00016588153324598076, + "loss": 0.2102, + "step": 32810 + }, + { + "epoch": 0.5119485867598428, + "grad_norm": 0.721671998500824, + "learning_rate": 0.00016587113412782598, + "loss": 0.2965, + "step": 32820 + }, + { + "epoch": 0.5121045735321644, + "grad_norm": 0.6120114326477051, + "learning_rate": 0.0001658607350096712, + "loss": 0.2661, + "step": 32830 + }, + { + "epoch": 0.5122605603044862, + "grad_norm": 0.29123347997665405, + "learning_rate": 0.00016585033589151642, + "loss": 0.2455, + "step": 32840 + }, + { + "epoch": 0.5124165470768078, + "grad_norm": 2.7300868034362793, + "learning_rate": 0.00016583993677336164, + "loss": 0.2049, + "step": 32850 + }, + { + "epoch": 0.5125725338491296, + "grad_norm": 1.801928997039795, + "learning_rate": 0.00016582953765520686, + "loss": 0.1725, + "step": 32860 + }, + { + "epoch": 0.5127285206214514, + "grad_norm": 0.6406471729278564, + "learning_rate": 0.00016581913853705208, + "loss": 0.164, + "step": 32870 + }, + { + "epoch": 0.512884507393773, + "grad_norm": 0.9627325534820557, + "learning_rate": 0.0001658087394188973, + "loss": 0.1728, + "step": 32880 + }, + { + "epoch": 0.5130404941660948, + "grad_norm": 1.5446962118148804, + "learning_rate": 0.00016579834030074251, + "loss": 0.2345, + "step": 32890 + }, + { + "epoch": 0.5131964809384164, + "grad_norm": 0.28092706203460693, + "learning_rate": 0.0001657879411825877, + "loss": 0.2124, + "step": 32900 + }, + { + "epoch": 0.5133524677107382, + "grad_norm": 1.7751421928405762, + "learning_rate": 0.00016577754206443295, + "loss": 0.2367, + "step": 32910 + }, + { + "epoch": 0.5135084544830598, + "grad_norm": 1.6964634656906128, + "learning_rate": 0.00016576714294627815, + "loss": 0.2361, + "step": 32920 + }, + { + "epoch": 0.5136644412553816, + "grad_norm": 0.06759823113679886, + "learning_rate": 0.0001657567438281234, + "loss": 0.2329, + "step": 32930 + }, + { + "epoch": 0.5138204280277032, + "grad_norm": 2.2519280910491943, + "learning_rate": 0.00016574634470996859, + "loss": 0.1079, + "step": 32940 + }, + { + "epoch": 0.513976414800025, + "grad_norm": 3.167576551437378, + "learning_rate": 0.00016573594559181383, + "loss": 0.1764, + "step": 32950 + }, + { + "epoch": 0.5141324015723466, + "grad_norm": 1.6734397411346436, + "learning_rate": 0.00016572554647365903, + "loss": 0.2878, + "step": 32960 + }, + { + "epoch": 0.5142883883446684, + "grad_norm": 0.35627374053001404, + "learning_rate": 0.00016571514735550427, + "loss": 0.2337, + "step": 32970 + }, + { + "epoch": 0.51444437511699, + "grad_norm": 1.6321039199829102, + "learning_rate": 0.00016570474823734946, + "loss": 0.2551, + "step": 32980 + }, + { + "epoch": 0.5146003618893118, + "grad_norm": 1.352982997894287, + "learning_rate": 0.0001656943491191947, + "loss": 0.4467, + "step": 32990 + }, + { + "epoch": 0.5147563486616334, + "grad_norm": 0.3558129072189331, + "learning_rate": 0.0001656839500010399, + "loss": 0.1913, + "step": 33000 + }, + { + "epoch": 0.5149123354339552, + "grad_norm": 0.041972286999225616, + "learning_rate": 0.00016567355088288515, + "loss": 0.1704, + "step": 33010 + }, + { + "epoch": 0.515068322206277, + "grad_norm": 1.4168939590454102, + "learning_rate": 0.00016566315176473034, + "loss": 0.4866, + "step": 33020 + }, + { + "epoch": 0.5152243089785986, + "grad_norm": 3.0042929649353027, + "learning_rate": 0.0001656527526465756, + "loss": 0.2473, + "step": 33030 + }, + { + "epoch": 0.5153802957509204, + "grad_norm": 1.027274489402771, + "learning_rate": 0.00016564235352842078, + "loss": 0.436, + "step": 33040 + }, + { + "epoch": 0.515536282523242, + "grad_norm": 1.2330195903778076, + "learning_rate": 0.00016563195441026603, + "loss": 0.4615, + "step": 33050 + }, + { + "epoch": 0.5156922692955638, + "grad_norm": 2.7145066261291504, + "learning_rate": 0.00016562155529211122, + "loss": 0.278, + "step": 33060 + }, + { + "epoch": 0.5158482560678854, + "grad_norm": 1.1273893117904663, + "learning_rate": 0.00016561115617395647, + "loss": 0.3767, + "step": 33070 + }, + { + "epoch": 0.5160042428402072, + "grad_norm": 2.466265916824341, + "learning_rate": 0.00016560075705580166, + "loss": 0.1879, + "step": 33080 + }, + { + "epoch": 0.5161602296125288, + "grad_norm": 0.3759515583515167, + "learning_rate": 0.0001655903579376469, + "loss": 0.4004, + "step": 33090 + }, + { + "epoch": 0.5163162163848506, + "grad_norm": 1.8821336030960083, + "learning_rate": 0.0001655799588194921, + "loss": 0.1622, + "step": 33100 + }, + { + "epoch": 0.5164722031571722, + "grad_norm": 0.8162096738815308, + "learning_rate": 0.00016556955970133735, + "loss": 0.1341, + "step": 33110 + }, + { + "epoch": 0.516628189929494, + "grad_norm": 0.008498801849782467, + "learning_rate": 0.00016555916058318254, + "loss": 0.2546, + "step": 33120 + }, + { + "epoch": 0.5167841767018156, + "grad_norm": 0.17536675930023193, + "learning_rate": 0.00016554876146502779, + "loss": 0.2934, + "step": 33130 + }, + { + "epoch": 0.5169401634741374, + "grad_norm": 3.539898633956909, + "learning_rate": 0.00016553836234687298, + "loss": 0.4959, + "step": 33140 + }, + { + "epoch": 0.517096150246459, + "grad_norm": 2.190887212753296, + "learning_rate": 0.00016552796322871823, + "loss": 0.2914, + "step": 33150 + }, + { + "epoch": 0.5172521370187808, + "grad_norm": 1.6388943195343018, + "learning_rate": 0.00016551756411056342, + "loss": 0.1317, + "step": 33160 + }, + { + "epoch": 0.5174081237911026, + "grad_norm": 0.33063387870788574, + "learning_rate": 0.00016550716499240866, + "loss": 0.4258, + "step": 33170 + }, + { + "epoch": 0.5175641105634242, + "grad_norm": 0.5168288946151733, + "learning_rate": 0.00016549676587425386, + "loss": 0.2548, + "step": 33180 + }, + { + "epoch": 0.517720097335746, + "grad_norm": 3.9570281505584717, + "learning_rate": 0.0001654863667560991, + "loss": 0.3179, + "step": 33190 + }, + { + "epoch": 0.5178760841080676, + "grad_norm": 1.127255916595459, + "learning_rate": 0.0001654759676379443, + "loss": 0.233, + "step": 33200 + }, + { + "epoch": 0.5180320708803894, + "grad_norm": 0.36299410462379456, + "learning_rate": 0.00016546556851978954, + "loss": 0.4614, + "step": 33210 + }, + { + "epoch": 0.518188057652711, + "grad_norm": 2.914668321609497, + "learning_rate": 0.00016545516940163474, + "loss": 0.1603, + "step": 33220 + }, + { + "epoch": 0.5183440444250328, + "grad_norm": 6.991094589233398, + "learning_rate": 0.00016544477028347998, + "loss": 0.2259, + "step": 33230 + }, + { + "epoch": 0.5185000311973544, + "grad_norm": 2.0594241619110107, + "learning_rate": 0.00016543437116532517, + "loss": 0.1885, + "step": 33240 + }, + { + "epoch": 0.5186560179696762, + "grad_norm": 0.5594693422317505, + "learning_rate": 0.00016542397204717042, + "loss": 0.4333, + "step": 33250 + }, + { + "epoch": 0.5188120047419978, + "grad_norm": 4.314854621887207, + "learning_rate": 0.00016541357292901561, + "loss": 0.2103, + "step": 33260 + }, + { + "epoch": 0.5189679915143196, + "grad_norm": 0.5609000325202942, + "learning_rate": 0.00016540317381086086, + "loss": 0.2498, + "step": 33270 + }, + { + "epoch": 0.5191239782866413, + "grad_norm": 3.680088758468628, + "learning_rate": 0.00016539277469270605, + "loss": 0.3465, + "step": 33280 + }, + { + "epoch": 0.519279965058963, + "grad_norm": 1.4993672370910645, + "learning_rate": 0.0001653823755745513, + "loss": 0.3272, + "step": 33290 + }, + { + "epoch": 0.5194359518312847, + "grad_norm": 0.18005084991455078, + "learning_rate": 0.0001653719764563965, + "loss": 0.1706, + "step": 33300 + }, + { + "epoch": 0.5195919386036064, + "grad_norm": 0.5105524659156799, + "learning_rate": 0.00016536157733824174, + "loss": 0.0937, + "step": 33310 + }, + { + "epoch": 0.5197479253759281, + "grad_norm": 1.2806589603424072, + "learning_rate": 0.00016535117822008693, + "loss": 0.1578, + "step": 33320 + }, + { + "epoch": 0.5199039121482498, + "grad_norm": 0.22015990316867828, + "learning_rate": 0.00016534077910193218, + "loss": 0.2081, + "step": 33330 + }, + { + "epoch": 0.5200598989205716, + "grad_norm": 0.34638991951942444, + "learning_rate": 0.00016533037998377737, + "loss": 0.2389, + "step": 33340 + }, + { + "epoch": 0.5202158856928932, + "grad_norm": 0.8709302544593811, + "learning_rate": 0.0001653199808656226, + "loss": 0.2112, + "step": 33350 + }, + { + "epoch": 0.520371872465215, + "grad_norm": 1.2309521436691284, + "learning_rate": 0.0001653095817474678, + "loss": 0.3448, + "step": 33360 + }, + { + "epoch": 0.5205278592375366, + "grad_norm": 1.3155115842819214, + "learning_rate": 0.00016529918262931303, + "loss": 0.1572, + "step": 33370 + }, + { + "epoch": 0.5206838460098584, + "grad_norm": 0.358751118183136, + "learning_rate": 0.00016528878351115825, + "loss": 0.3011, + "step": 33380 + }, + { + "epoch": 0.52083983278218, + "grad_norm": 3.800628662109375, + "learning_rate": 0.00016527838439300347, + "loss": 0.458, + "step": 33390 + }, + { + "epoch": 0.5209958195545018, + "grad_norm": 1.1253124475479126, + "learning_rate": 0.0001652679852748487, + "loss": 0.1238, + "step": 33400 + }, + { + "epoch": 0.5211518063268235, + "grad_norm": 0.46979108452796936, + "learning_rate": 0.0001652575861566939, + "loss": 0.0745, + "step": 33410 + }, + { + "epoch": 0.5213077930991452, + "grad_norm": 7.921492576599121, + "learning_rate": 0.00016524718703853913, + "loss": 0.4004, + "step": 33420 + }, + { + "epoch": 0.5214637798714669, + "grad_norm": 1.643127679824829, + "learning_rate": 0.00016523678792038435, + "loss": 0.2858, + "step": 33430 + }, + { + "epoch": 0.5216197666437886, + "grad_norm": 7.096631050109863, + "learning_rate": 0.0001652263888022296, + "loss": 0.4735, + "step": 33440 + }, + { + "epoch": 0.5217757534161103, + "grad_norm": 0.7280219197273254, + "learning_rate": 0.0001652159896840748, + "loss": 0.1841, + "step": 33450 + }, + { + "epoch": 0.521931740188432, + "grad_norm": 0.9415522813796997, + "learning_rate": 0.00016520559056592003, + "loss": 0.2169, + "step": 33460 + }, + { + "epoch": 0.5220877269607537, + "grad_norm": 2.264862060546875, + "learning_rate": 0.00016519519144776523, + "loss": 0.1902, + "step": 33470 + }, + { + "epoch": 0.5222437137330754, + "grad_norm": 2.4566822052001953, + "learning_rate": 0.00016518479232961047, + "loss": 0.3329, + "step": 33480 + }, + { + "epoch": 0.5223997005053972, + "grad_norm": 0.7330378890037537, + "learning_rate": 0.00016517439321145567, + "loss": 0.2341, + "step": 33490 + }, + { + "epoch": 0.5225556872777188, + "grad_norm": 1.6409542560577393, + "learning_rate": 0.0001651639940933009, + "loss": 0.3077, + "step": 33500 + }, + { + "epoch": 0.5227116740500406, + "grad_norm": 1.0647432804107666, + "learning_rate": 0.0001651535949751461, + "loss": 0.1935, + "step": 33510 + }, + { + "epoch": 0.5228676608223622, + "grad_norm": 0.507420539855957, + "learning_rate": 0.00016514319585699135, + "loss": 0.1314, + "step": 33520 + }, + { + "epoch": 0.523023647594684, + "grad_norm": 1.0064164400100708, + "learning_rate": 0.00016513279673883654, + "loss": 0.344, + "step": 33530 + }, + { + "epoch": 0.5231796343670057, + "grad_norm": 1.1936209201812744, + "learning_rate": 0.0001651223976206818, + "loss": 0.1948, + "step": 33540 + }, + { + "epoch": 0.5233356211393274, + "grad_norm": 0.04992926865816116, + "learning_rate": 0.00016511199850252698, + "loss": 0.1363, + "step": 33550 + }, + { + "epoch": 0.5234916079116491, + "grad_norm": 1.8256275653839111, + "learning_rate": 0.00016510159938437223, + "loss": 0.443, + "step": 33560 + }, + { + "epoch": 0.5236475946839708, + "grad_norm": 4.914908409118652, + "learning_rate": 0.00016509120026621742, + "loss": 0.4654, + "step": 33570 + }, + { + "epoch": 0.5238035814562925, + "grad_norm": 1.1625486612319946, + "learning_rate": 0.00016508080114806267, + "loss": 0.1974, + "step": 33580 + }, + { + "epoch": 0.5239595682286142, + "grad_norm": 6.6192522048950195, + "learning_rate": 0.00016507040202990786, + "loss": 0.2131, + "step": 33590 + }, + { + "epoch": 0.5241155550009359, + "grad_norm": 0.7369065880775452, + "learning_rate": 0.0001650600029117531, + "loss": 0.2371, + "step": 33600 + }, + { + "epoch": 0.5242715417732576, + "grad_norm": 1.5238152742385864, + "learning_rate": 0.0001650496037935983, + "loss": 0.1557, + "step": 33610 + }, + { + "epoch": 0.5244275285455793, + "grad_norm": 1.0418007373809814, + "learning_rate": 0.00016503920467544355, + "loss": 0.1878, + "step": 33620 + }, + { + "epoch": 0.524583515317901, + "grad_norm": 0.790117084980011, + "learning_rate": 0.00016502880555728874, + "loss": 0.2195, + "step": 33630 + }, + { + "epoch": 0.5247395020902228, + "grad_norm": 1.6712257862091064, + "learning_rate": 0.000165018406439134, + "loss": 0.1602, + "step": 33640 + }, + { + "epoch": 0.5248954888625444, + "grad_norm": 0.19236230850219727, + "learning_rate": 0.00016500800732097918, + "loss": 0.2526, + "step": 33650 + }, + { + "epoch": 0.5250514756348662, + "grad_norm": 1.3519701957702637, + "learning_rate": 0.00016499760820282443, + "loss": 0.4686, + "step": 33660 + }, + { + "epoch": 0.5252074624071879, + "grad_norm": 1.694342851638794, + "learning_rate": 0.00016498720908466962, + "loss": 0.1859, + "step": 33670 + }, + { + "epoch": 0.5253634491795096, + "grad_norm": 5.225239276885986, + "learning_rate": 0.00016497680996651487, + "loss": 0.2944, + "step": 33680 + }, + { + "epoch": 0.5255194359518313, + "grad_norm": 2.0208842754364014, + "learning_rate": 0.00016496641084836006, + "loss": 0.2421, + "step": 33690 + }, + { + "epoch": 0.525675422724153, + "grad_norm": 0.7954996228218079, + "learning_rate": 0.0001649560117302053, + "loss": 0.2853, + "step": 33700 + }, + { + "epoch": 0.5258314094964747, + "grad_norm": 2.296086072921753, + "learning_rate": 0.0001649456126120505, + "loss": 0.1333, + "step": 33710 + }, + { + "epoch": 0.5259873962687964, + "grad_norm": 1.1779128313064575, + "learning_rate": 0.00016493521349389574, + "loss": 0.1066, + "step": 33720 + }, + { + "epoch": 0.5261433830411181, + "grad_norm": 0.1756065934896469, + "learning_rate": 0.00016492481437574094, + "loss": 0.1352, + "step": 33730 + }, + { + "epoch": 0.5262993698134398, + "grad_norm": 0.13100725412368774, + "learning_rate": 0.00016491441525758618, + "loss": 0.2399, + "step": 33740 + }, + { + "epoch": 0.5264553565857615, + "grad_norm": 5.532008171081543, + "learning_rate": 0.00016490401613943138, + "loss": 0.2896, + "step": 33750 + }, + { + "epoch": 0.5266113433580832, + "grad_norm": 1.319886565208435, + "learning_rate": 0.00016489361702127662, + "loss": 0.3275, + "step": 33760 + }, + { + "epoch": 0.5267673301304049, + "grad_norm": 1.5550974607467651, + "learning_rate": 0.00016488321790312182, + "loss": 0.2677, + "step": 33770 + }, + { + "epoch": 0.5269233169027266, + "grad_norm": 1.8936737775802612, + "learning_rate": 0.00016487281878496706, + "loss": 0.1955, + "step": 33780 + }, + { + "epoch": 0.5270793036750484, + "grad_norm": 0.3653401732444763, + "learning_rate": 0.00016486241966681226, + "loss": 0.0723, + "step": 33790 + }, + { + "epoch": 0.52723529044737, + "grad_norm": 2.861341714859009, + "learning_rate": 0.00016485202054865747, + "loss": 0.2412, + "step": 33800 + }, + { + "epoch": 0.5273912772196918, + "grad_norm": 1.5291428565979004, + "learning_rate": 0.0001648416214305027, + "loss": 0.0871, + "step": 33810 + }, + { + "epoch": 0.5275472639920135, + "grad_norm": 1.0372581481933594, + "learning_rate": 0.00016483122231234791, + "loss": 0.4705, + "step": 33820 + }, + { + "epoch": 0.5277032507643352, + "grad_norm": 1.1943141222000122, + "learning_rate": 0.00016482082319419313, + "loss": 0.2848, + "step": 33830 + }, + { + "epoch": 0.5278592375366569, + "grad_norm": 1.9008225202560425, + "learning_rate": 0.00016481042407603835, + "loss": 0.139, + "step": 33840 + }, + { + "epoch": 0.5280152243089786, + "grad_norm": 2.132089138031006, + "learning_rate": 0.00016480002495788357, + "loss": 0.2119, + "step": 33850 + }, + { + "epoch": 0.5281712110813003, + "grad_norm": 0.24524426460266113, + "learning_rate": 0.0001647896258397288, + "loss": 0.1643, + "step": 33860 + }, + { + "epoch": 0.528327197853622, + "grad_norm": 1.6469637155532837, + "learning_rate": 0.000164779226721574, + "loss": 0.363, + "step": 33870 + }, + { + "epoch": 0.5284831846259437, + "grad_norm": 0.8767328858375549, + "learning_rate": 0.00016476882760341923, + "loss": 0.2632, + "step": 33880 + }, + { + "epoch": 0.5286391713982654, + "grad_norm": 0.06347586214542389, + "learning_rate": 0.00016475842848526445, + "loss": 0.3204, + "step": 33890 + }, + { + "epoch": 0.5287951581705871, + "grad_norm": 0.09782540798187256, + "learning_rate": 0.00016474802936710967, + "loss": 0.2211, + "step": 33900 + }, + { + "epoch": 0.5289511449429088, + "grad_norm": 3.2998859882354736, + "learning_rate": 0.0001647376302489549, + "loss": 0.265, + "step": 33910 + }, + { + "epoch": 0.5291071317152305, + "grad_norm": 0.43594226241111755, + "learning_rate": 0.0001647272311308001, + "loss": 0.194, + "step": 33920 + }, + { + "epoch": 0.5292631184875523, + "grad_norm": 1.5166605710983276, + "learning_rate": 0.00016471683201264533, + "loss": 0.2675, + "step": 33930 + }, + { + "epoch": 0.529419105259874, + "grad_norm": 0.6056640148162842, + "learning_rate": 0.00016470643289449055, + "loss": 0.0576, + "step": 33940 + }, + { + "epoch": 0.5295750920321957, + "grad_norm": 0.25410348176956177, + "learning_rate": 0.00016469603377633577, + "loss": 0.1124, + "step": 33950 + }, + { + "epoch": 0.5297310788045174, + "grad_norm": 1.770642876625061, + "learning_rate": 0.000164685634658181, + "loss": 0.3295, + "step": 33960 + }, + { + "epoch": 0.5298870655768391, + "grad_norm": 0.0607205331325531, + "learning_rate": 0.0001646752355400262, + "loss": 0.2369, + "step": 33970 + }, + { + "epoch": 0.5300430523491608, + "grad_norm": 0.5557095408439636, + "learning_rate": 0.00016466483642187143, + "loss": 0.1681, + "step": 33980 + }, + { + "epoch": 0.5301990391214825, + "grad_norm": 0.5192957520484924, + "learning_rate": 0.00016465443730371665, + "loss": 0.2649, + "step": 33990 + }, + { + "epoch": 0.5303550258938042, + "grad_norm": 0.04804835096001625, + "learning_rate": 0.00016464403818556187, + "loss": 0.2639, + "step": 34000 + }, + { + "epoch": 0.5305110126661259, + "grad_norm": 0.02673129364848137, + "learning_rate": 0.0001646336390674071, + "loss": 0.169, + "step": 34010 + }, + { + "epoch": 0.5306669994384476, + "grad_norm": 1.0084244012832642, + "learning_rate": 0.0001646232399492523, + "loss": 0.353, + "step": 34020 + }, + { + "epoch": 0.5308229862107693, + "grad_norm": 2.2202091217041016, + "learning_rate": 0.00016461284083109753, + "loss": 0.3206, + "step": 34030 + }, + { + "epoch": 0.530978972983091, + "grad_norm": 0.5573744773864746, + "learning_rate": 0.00016460244171294275, + "loss": 0.231, + "step": 34040 + }, + { + "epoch": 0.5311349597554127, + "grad_norm": 0.39700084924697876, + "learning_rate": 0.00016459204259478797, + "loss": 0.1412, + "step": 34050 + }, + { + "epoch": 0.5312909465277345, + "grad_norm": 2.582963228225708, + "learning_rate": 0.00016458164347663319, + "loss": 0.3334, + "step": 34060 + }, + { + "epoch": 0.5314469333000561, + "grad_norm": 0.22781169414520264, + "learning_rate": 0.0001645712443584784, + "loss": 0.2767, + "step": 34070 + }, + { + "epoch": 0.5316029200723779, + "grad_norm": 1.2051042318344116, + "learning_rate": 0.00016456084524032362, + "loss": 0.2045, + "step": 34080 + }, + { + "epoch": 0.5317589068446996, + "grad_norm": 0.42760100960731506, + "learning_rate": 0.00016455044612216884, + "loss": 0.1151, + "step": 34090 + }, + { + "epoch": 0.5319148936170213, + "grad_norm": 0.14440476894378662, + "learning_rate": 0.00016454004700401406, + "loss": 0.2386, + "step": 34100 + }, + { + "epoch": 0.532070880389343, + "grad_norm": 2.2777981758117676, + "learning_rate": 0.00016452964788585928, + "loss": 0.2009, + "step": 34110 + }, + { + "epoch": 0.5322268671616647, + "grad_norm": 0.9206979274749756, + "learning_rate": 0.0001645192487677045, + "loss": 0.2745, + "step": 34120 + }, + { + "epoch": 0.5323828539339864, + "grad_norm": 1.6947574615478516, + "learning_rate": 0.00016450884964954972, + "loss": 0.2584, + "step": 34130 + }, + { + "epoch": 0.5325388407063081, + "grad_norm": 0.401444673538208, + "learning_rate": 0.00016449845053139494, + "loss": 0.2218, + "step": 34140 + }, + { + "epoch": 0.5326948274786298, + "grad_norm": 0.08261553198099136, + "learning_rate": 0.00016448805141324016, + "loss": 0.2775, + "step": 34150 + }, + { + "epoch": 0.5328508142509515, + "grad_norm": 0.1017974391579628, + "learning_rate": 0.00016447765229508538, + "loss": 0.2095, + "step": 34160 + }, + { + "epoch": 0.5330068010232732, + "grad_norm": 1.3759571313858032, + "learning_rate": 0.0001644672531769306, + "loss": 0.2643, + "step": 34170 + }, + { + "epoch": 0.5331627877955949, + "grad_norm": 1.2654389142990112, + "learning_rate": 0.00016445685405877582, + "loss": 0.2949, + "step": 34180 + }, + { + "epoch": 0.5333187745679167, + "grad_norm": 1.5481843948364258, + "learning_rate": 0.00016444645494062104, + "loss": 0.1264, + "step": 34190 + }, + { + "epoch": 0.5334747613402383, + "grad_norm": 1.8094528913497925, + "learning_rate": 0.00016443605582246626, + "loss": 0.2727, + "step": 34200 + }, + { + "epoch": 0.5336307481125601, + "grad_norm": 2.224538564682007, + "learning_rate": 0.00016442565670431148, + "loss": 0.3096, + "step": 34210 + }, + { + "epoch": 0.5337867348848817, + "grad_norm": 0.6375226974487305, + "learning_rate": 0.0001644152575861567, + "loss": 0.2251, + "step": 34220 + }, + { + "epoch": 0.5339427216572035, + "grad_norm": 3.727106809616089, + "learning_rate": 0.00016440485846800192, + "loss": 0.4374, + "step": 34230 + }, + { + "epoch": 0.5340987084295252, + "grad_norm": 0.13345426321029663, + "learning_rate": 0.00016439445934984714, + "loss": 0.2011, + "step": 34240 + }, + { + "epoch": 0.5342546952018469, + "grad_norm": 2.1658668518066406, + "learning_rate": 0.00016438406023169236, + "loss": 0.2457, + "step": 34250 + }, + { + "epoch": 0.5344106819741686, + "grad_norm": 19.238407135009766, + "learning_rate": 0.00016437366111353758, + "loss": 0.2756, + "step": 34260 + }, + { + "epoch": 0.5345666687464903, + "grad_norm": 1.0292778015136719, + "learning_rate": 0.0001643632619953828, + "loss": 0.1646, + "step": 34270 + }, + { + "epoch": 0.534722655518812, + "grad_norm": 0.9372987747192383, + "learning_rate": 0.00016435286287722802, + "loss": 0.2762, + "step": 34280 + }, + { + "epoch": 0.5348786422911337, + "grad_norm": 0.3918002247810364, + "learning_rate": 0.00016434246375907324, + "loss": 0.1406, + "step": 34290 + }, + { + "epoch": 0.5350346290634554, + "grad_norm": 1.3518732786178589, + "learning_rate": 0.00016433206464091846, + "loss": 0.2992, + "step": 34300 + }, + { + "epoch": 0.5351906158357771, + "grad_norm": 0.73117595911026, + "learning_rate": 0.00016432166552276368, + "loss": 0.1172, + "step": 34310 + }, + { + "epoch": 0.5353466026080989, + "grad_norm": 0.06655958294868469, + "learning_rate": 0.0001643112664046089, + "loss": 0.1332, + "step": 34320 + }, + { + "epoch": 0.5355025893804205, + "grad_norm": 3.2587168216705322, + "learning_rate": 0.00016430086728645412, + "loss": 0.2052, + "step": 34330 + }, + { + "epoch": 0.5356585761527423, + "grad_norm": 2.6668431758880615, + "learning_rate": 0.00016429046816829934, + "loss": 0.1295, + "step": 34340 + }, + { + "epoch": 0.5358145629250639, + "grad_norm": 1.316080927848816, + "learning_rate": 0.00016428006905014456, + "loss": 0.2072, + "step": 34350 + }, + { + "epoch": 0.5359705496973857, + "grad_norm": 5.144528388977051, + "learning_rate": 0.00016426966993198977, + "loss": 0.1934, + "step": 34360 + }, + { + "epoch": 0.5361265364697073, + "grad_norm": 1.3529599905014038, + "learning_rate": 0.000164259270813835, + "loss": 0.2077, + "step": 34370 + }, + { + "epoch": 0.5362825232420291, + "grad_norm": 0.9132925868034363, + "learning_rate": 0.00016424887169568021, + "loss": 0.2404, + "step": 34380 + }, + { + "epoch": 0.5364385100143508, + "grad_norm": 1.520033597946167, + "learning_rate": 0.00016423847257752543, + "loss": 0.1765, + "step": 34390 + }, + { + "epoch": 0.5365944967866725, + "grad_norm": 2.343975305557251, + "learning_rate": 0.00016422807345937065, + "loss": 0.1959, + "step": 34400 + }, + { + "epoch": 0.5367504835589942, + "grad_norm": 0.5855118632316589, + "learning_rate": 0.00016421767434121587, + "loss": 0.0848, + "step": 34410 + }, + { + "epoch": 0.5369064703313159, + "grad_norm": 1.9108648300170898, + "learning_rate": 0.0001642072752230611, + "loss": 0.2406, + "step": 34420 + }, + { + "epoch": 0.5370624571036376, + "grad_norm": 1.0323792695999146, + "learning_rate": 0.0001641968761049063, + "loss": 0.2298, + "step": 34430 + }, + { + "epoch": 0.5372184438759593, + "grad_norm": 0.17435620725154877, + "learning_rate": 0.00016418647698675153, + "loss": 0.1396, + "step": 34440 + }, + { + "epoch": 0.537374430648281, + "grad_norm": 0.3526577353477478, + "learning_rate": 0.00016417607786859675, + "loss": 0.172, + "step": 34450 + }, + { + "epoch": 0.5375304174206027, + "grad_norm": 2.73685884475708, + "learning_rate": 0.00016416567875044197, + "loss": 0.1167, + "step": 34460 + }, + { + "epoch": 0.5376864041929245, + "grad_norm": 2.765693426132202, + "learning_rate": 0.0001641552796322872, + "loss": 0.1356, + "step": 34470 + }, + { + "epoch": 0.5378423909652461, + "grad_norm": 0.47628021240234375, + "learning_rate": 0.0001641448805141324, + "loss": 0.3104, + "step": 34480 + }, + { + "epoch": 0.5379983777375679, + "grad_norm": 1.1307590007781982, + "learning_rate": 0.00016413448139597763, + "loss": 0.3247, + "step": 34490 + }, + { + "epoch": 0.5381543645098895, + "grad_norm": 0.07136381417512894, + "learning_rate": 0.00016412408227782285, + "loss": 0.319, + "step": 34500 + }, + { + "epoch": 0.5383103512822113, + "grad_norm": 1.7850221395492554, + "learning_rate": 0.00016411368315966807, + "loss": 0.2515, + "step": 34510 + }, + { + "epoch": 0.5384663380545329, + "grad_norm": 1.3870742321014404, + "learning_rate": 0.0001641032840415133, + "loss": 0.256, + "step": 34520 + }, + { + "epoch": 0.5386223248268547, + "grad_norm": 1.1329221725463867, + "learning_rate": 0.0001640928849233585, + "loss": 0.2255, + "step": 34530 + }, + { + "epoch": 0.5387783115991764, + "grad_norm": 0.23236137628555298, + "learning_rate": 0.00016408248580520373, + "loss": 0.4053, + "step": 34540 + }, + { + "epoch": 0.5389342983714981, + "grad_norm": 3.350924253463745, + "learning_rate": 0.00016407208668704895, + "loss": 0.2816, + "step": 34550 + }, + { + "epoch": 0.5390902851438198, + "grad_norm": 1.7459170818328857, + "learning_rate": 0.00016406168756889417, + "loss": 0.0834, + "step": 34560 + }, + { + "epoch": 0.5392462719161415, + "grad_norm": 0.89351487159729, + "learning_rate": 0.0001640512884507394, + "loss": 0.1146, + "step": 34570 + }, + { + "epoch": 0.5394022586884633, + "grad_norm": 0.49410831928253174, + "learning_rate": 0.0001640408893325846, + "loss": 0.2319, + "step": 34580 + }, + { + "epoch": 0.5395582454607849, + "grad_norm": 0.009971237741410732, + "learning_rate": 0.00016403049021442983, + "loss": 0.1674, + "step": 34590 + }, + { + "epoch": 0.5397142322331067, + "grad_norm": 1.420511245727539, + "learning_rate": 0.00016402009109627505, + "loss": 0.1469, + "step": 34600 + }, + { + "epoch": 0.5398702190054283, + "grad_norm": 2.5686769485473633, + "learning_rate": 0.00016400969197812027, + "loss": 0.2607, + "step": 34610 + }, + { + "epoch": 0.5400262057777501, + "grad_norm": 2.5808913707733154, + "learning_rate": 0.00016399929285996549, + "loss": 0.4412, + "step": 34620 + }, + { + "epoch": 0.5401821925500717, + "grad_norm": 0.2778591811656952, + "learning_rate": 0.0001639888937418107, + "loss": 0.1563, + "step": 34630 + }, + { + "epoch": 0.5403381793223935, + "grad_norm": 1.6448099613189697, + "learning_rate": 0.00016397849462365592, + "loss": 0.2484, + "step": 34640 + }, + { + "epoch": 0.5404941660947151, + "grad_norm": 0.003213417250663042, + "learning_rate": 0.00016396809550550114, + "loss": 0.1582, + "step": 34650 + }, + { + "epoch": 0.5406501528670369, + "grad_norm": 2.3433990478515625, + "learning_rate": 0.00016395769638734636, + "loss": 0.2774, + "step": 34660 + }, + { + "epoch": 0.5408061396393585, + "grad_norm": 0.09544426947832108, + "learning_rate": 0.00016394729726919158, + "loss": 0.1242, + "step": 34670 + }, + { + "epoch": 0.5409621264116803, + "grad_norm": 1.8934235572814941, + "learning_rate": 0.0001639368981510368, + "loss": 0.1636, + "step": 34680 + }, + { + "epoch": 0.541118113184002, + "grad_norm": 2.8528554439544678, + "learning_rate": 0.00016392649903288202, + "loss": 0.2524, + "step": 34690 + }, + { + "epoch": 0.5412740999563237, + "grad_norm": 2.837836980819702, + "learning_rate": 0.00016391609991472722, + "loss": 0.1808, + "step": 34700 + }, + { + "epoch": 0.5414300867286455, + "grad_norm": 4.349740028381348, + "learning_rate": 0.00016390570079657246, + "loss": 0.12, + "step": 34710 + }, + { + "epoch": 0.5415860735009671, + "grad_norm": 0.6011945605278015, + "learning_rate": 0.00016389530167841765, + "loss": 0.2865, + "step": 34720 + }, + { + "epoch": 0.5417420602732889, + "grad_norm": 1.2079488039016724, + "learning_rate": 0.0001638849025602629, + "loss": 0.125, + "step": 34730 + }, + { + "epoch": 0.5418980470456105, + "grad_norm": 1.9549164772033691, + "learning_rate": 0.0001638745034421081, + "loss": 0.3285, + "step": 34740 + }, + { + "epoch": 0.5420540338179323, + "grad_norm": 0.6969407200813293, + "learning_rate": 0.00016386410432395334, + "loss": 0.3267, + "step": 34750 + }, + { + "epoch": 0.5422100205902539, + "grad_norm": 1.0561965703964233, + "learning_rate": 0.00016385370520579853, + "loss": 0.1463, + "step": 34760 + }, + { + "epoch": 0.5423660073625757, + "grad_norm": 1.8208931684494019, + "learning_rate": 0.00016384330608764378, + "loss": 0.1713, + "step": 34770 + }, + { + "epoch": 0.5425219941348973, + "grad_norm": 0.6349910497665405, + "learning_rate": 0.000163832906969489, + "loss": 0.2244, + "step": 34780 + }, + { + "epoch": 0.5426779809072191, + "grad_norm": 1.1976515054702759, + "learning_rate": 0.00016382250785133422, + "loss": 0.3407, + "step": 34790 + }, + { + "epoch": 0.5428339676795407, + "grad_norm": 0.5970319509506226, + "learning_rate": 0.00016381210873317944, + "loss": 0.1419, + "step": 34800 + }, + { + "epoch": 0.5429899544518625, + "grad_norm": 0.14623159170150757, + "learning_rate": 0.00016380170961502466, + "loss": 0.14, + "step": 34810 + }, + { + "epoch": 0.5431459412241841, + "grad_norm": 1.4332351684570312, + "learning_rate": 0.00016379131049686988, + "loss": 0.1093, + "step": 34820 + }, + { + "epoch": 0.5433019279965059, + "grad_norm": 1.1443063020706177, + "learning_rate": 0.0001637809113787151, + "loss": 0.3676, + "step": 34830 + }, + { + "epoch": 0.5434579147688277, + "grad_norm": 1.3897461891174316, + "learning_rate": 0.00016377051226056032, + "loss": 0.3713, + "step": 34840 + }, + { + "epoch": 0.5436139015411493, + "grad_norm": 1.7773199081420898, + "learning_rate": 0.00016376011314240554, + "loss": 0.2249, + "step": 34850 + }, + { + "epoch": 0.5437698883134711, + "grad_norm": 0.8524389266967773, + "learning_rate": 0.00016374971402425076, + "loss": 0.2158, + "step": 34860 + }, + { + "epoch": 0.5439258750857927, + "grad_norm": 0.5572389960289001, + "learning_rate": 0.00016373931490609598, + "loss": 0.215, + "step": 34870 + }, + { + "epoch": 0.5440818618581145, + "grad_norm": 3.0238709449768066, + "learning_rate": 0.0001637289157879412, + "loss": 0.3765, + "step": 34880 + }, + { + "epoch": 0.5442378486304361, + "grad_norm": 1.1655421257019043, + "learning_rate": 0.00016371851666978642, + "loss": 0.2491, + "step": 34890 + }, + { + "epoch": 0.5443938354027579, + "grad_norm": 0.49492090940475464, + "learning_rate": 0.00016370811755163164, + "loss": 0.1287, + "step": 34900 + }, + { + "epoch": 0.5445498221750795, + "grad_norm": 0.2732921540737152, + "learning_rate": 0.00016369771843347685, + "loss": 0.2154, + "step": 34910 + }, + { + "epoch": 0.5447058089474013, + "grad_norm": 2.5807769298553467, + "learning_rate": 0.00016368731931532207, + "loss": 0.2032, + "step": 34920 + }, + { + "epoch": 0.5448617957197229, + "grad_norm": 0.5044315457344055, + "learning_rate": 0.0001636769201971673, + "loss": 0.2595, + "step": 34930 + }, + { + "epoch": 0.5450177824920447, + "grad_norm": 2.1248972415924072, + "learning_rate": 0.00016366652107901251, + "loss": 0.5651, + "step": 34940 + }, + { + "epoch": 0.5451737692643663, + "grad_norm": 0.8391468524932861, + "learning_rate": 0.00016365612196085773, + "loss": 0.3028, + "step": 34950 + }, + { + "epoch": 0.5453297560366881, + "grad_norm": 0.36081477999687195, + "learning_rate": 0.00016364572284270295, + "loss": 0.1073, + "step": 34960 + }, + { + "epoch": 0.5454857428090097, + "grad_norm": 3.175804376602173, + "learning_rate": 0.00016363532372454817, + "loss": 0.1889, + "step": 34970 + }, + { + "epoch": 0.5456417295813315, + "grad_norm": 0.20647937059402466, + "learning_rate": 0.0001636249246063934, + "loss": 0.2416, + "step": 34980 + }, + { + "epoch": 0.5457977163536533, + "grad_norm": 2.3525943756103516, + "learning_rate": 0.0001636145254882386, + "loss": 0.3179, + "step": 34990 + }, + { + "epoch": 0.5459537031259749, + "grad_norm": 0.2641216814517975, + "learning_rate": 0.00016360412637008383, + "loss": 0.2157, + "step": 35000 + }, + { + "epoch": 0.5461096898982967, + "grad_norm": 0.09906073659658432, + "learning_rate": 0.00016359372725192905, + "loss": 0.2576, + "step": 35010 + }, + { + "epoch": 0.5462656766706183, + "grad_norm": 0.13971920311450958, + "learning_rate": 0.00016358332813377427, + "loss": 0.1298, + "step": 35020 + }, + { + "epoch": 0.5464216634429401, + "grad_norm": 1.4535638093948364, + "learning_rate": 0.0001635729290156195, + "loss": 0.2357, + "step": 35030 + }, + { + "epoch": 0.5465776502152617, + "grad_norm": 0.9107828736305237, + "learning_rate": 0.0001635625298974647, + "loss": 0.3049, + "step": 35040 + }, + { + "epoch": 0.5467336369875835, + "grad_norm": 0.0377386212348938, + "learning_rate": 0.00016355213077930993, + "loss": 0.2066, + "step": 35050 + }, + { + "epoch": 0.5468896237599051, + "grad_norm": 2.7800869941711426, + "learning_rate": 0.00016354173166115515, + "loss": 0.2974, + "step": 35060 + }, + { + "epoch": 0.5470456105322269, + "grad_norm": 1.6247998476028442, + "learning_rate": 0.00016353133254300037, + "loss": 0.3822, + "step": 35070 + }, + { + "epoch": 0.5472015973045485, + "grad_norm": 1.6479015350341797, + "learning_rate": 0.0001635209334248456, + "loss": 0.3401, + "step": 35080 + }, + { + "epoch": 0.5473575840768703, + "grad_norm": 0.1879737675189972, + "learning_rate": 0.0001635105343066908, + "loss": 0.1549, + "step": 35090 + }, + { + "epoch": 0.5475135708491919, + "grad_norm": 0.39355361461639404, + "learning_rate": 0.00016350013518853603, + "loss": 0.2032, + "step": 35100 + }, + { + "epoch": 0.5476695576215137, + "grad_norm": 2.9976983070373535, + "learning_rate": 0.00016348973607038125, + "loss": 0.3867, + "step": 35110 + }, + { + "epoch": 0.5478255443938354, + "grad_norm": 0.8242707252502441, + "learning_rate": 0.00016347933695222647, + "loss": 0.3601, + "step": 35120 + }, + { + "epoch": 0.5479815311661571, + "grad_norm": 0.874955415725708, + "learning_rate": 0.0001634689378340717, + "loss": 0.2709, + "step": 35130 + }, + { + "epoch": 0.5481375179384789, + "grad_norm": 0.05980971083045006, + "learning_rate": 0.0001634585387159169, + "loss": 0.2646, + "step": 35140 + }, + { + "epoch": 0.5482935047108005, + "grad_norm": 1.8437328338623047, + "learning_rate": 0.0001634481395977621, + "loss": 0.1965, + "step": 35150 + }, + { + "epoch": 0.5484494914831223, + "grad_norm": 0.4142405688762665, + "learning_rate": 0.00016343774047960735, + "loss": 0.1756, + "step": 35160 + }, + { + "epoch": 0.5486054782554439, + "grad_norm": 4.836324214935303, + "learning_rate": 0.00016342734136145254, + "loss": 0.2534, + "step": 35170 + }, + { + "epoch": 0.5487614650277657, + "grad_norm": 1.4933065176010132, + "learning_rate": 0.00016341694224329779, + "loss": 0.2355, + "step": 35180 + }, + { + "epoch": 0.5489174518000873, + "grad_norm": 1.9468894004821777, + "learning_rate": 0.00016340654312514298, + "loss": 0.3113, + "step": 35190 + }, + { + "epoch": 0.5490734385724091, + "grad_norm": 1.012710452079773, + "learning_rate": 0.00016339614400698822, + "loss": 0.2635, + "step": 35200 + }, + { + "epoch": 0.5492294253447307, + "grad_norm": 1.5551140308380127, + "learning_rate": 0.00016338574488883342, + "loss": 0.2951, + "step": 35210 + }, + { + "epoch": 0.5493854121170525, + "grad_norm": 1.8613696098327637, + "learning_rate": 0.00016337534577067866, + "loss": 0.2832, + "step": 35220 + }, + { + "epoch": 0.5495413988893741, + "grad_norm": 1.6748839616775513, + "learning_rate": 0.00016336494665252386, + "loss": 0.2266, + "step": 35230 + }, + { + "epoch": 0.5496973856616959, + "grad_norm": 0.4094032347202301, + "learning_rate": 0.0001633545475343691, + "loss": 0.3497, + "step": 35240 + }, + { + "epoch": 0.5498533724340176, + "grad_norm": 1.6414631605148315, + "learning_rate": 0.0001633441484162143, + "loss": 0.1763, + "step": 35250 + }, + { + "epoch": 0.5500093592063393, + "grad_norm": 1.6380645036697388, + "learning_rate": 0.00016333374929805954, + "loss": 0.1355, + "step": 35260 + }, + { + "epoch": 0.550165345978661, + "grad_norm": 0.7325630187988281, + "learning_rate": 0.00016332335017990473, + "loss": 0.3105, + "step": 35270 + }, + { + "epoch": 0.5503213327509827, + "grad_norm": 1.6015644073486328, + "learning_rate": 0.00016331295106174998, + "loss": 0.2311, + "step": 35280 + }, + { + "epoch": 0.5504773195233045, + "grad_norm": 0.9224210381507874, + "learning_rate": 0.00016330255194359517, + "loss": 0.2746, + "step": 35290 + }, + { + "epoch": 0.5506333062956261, + "grad_norm": 1.30025315284729, + "learning_rate": 0.00016329215282544042, + "loss": 0.2066, + "step": 35300 + }, + { + "epoch": 0.5507892930679479, + "grad_norm": 0.481125146150589, + "learning_rate": 0.0001632817537072856, + "loss": 0.1778, + "step": 35310 + }, + { + "epoch": 0.5509452798402695, + "grad_norm": 0.2709486782550812, + "learning_rate": 0.00016327135458913086, + "loss": 0.1131, + "step": 35320 + }, + { + "epoch": 0.5511012666125913, + "grad_norm": 0.6248563528060913, + "learning_rate": 0.00016326095547097605, + "loss": 0.4356, + "step": 35330 + }, + { + "epoch": 0.5512572533849129, + "grad_norm": 3.0947647094726562, + "learning_rate": 0.0001632505563528213, + "loss": 0.3072, + "step": 35340 + }, + { + "epoch": 0.5514132401572347, + "grad_norm": 1.009535312652588, + "learning_rate": 0.0001632401572346665, + "loss": 0.1808, + "step": 35350 + }, + { + "epoch": 0.5515692269295563, + "grad_norm": 0.042604975402355194, + "learning_rate": 0.00016322975811651174, + "loss": 0.1198, + "step": 35360 + }, + { + "epoch": 0.5517252137018781, + "grad_norm": 1.8652396202087402, + "learning_rate": 0.00016321935899835693, + "loss": 0.2176, + "step": 35370 + }, + { + "epoch": 0.5518812004741998, + "grad_norm": 4.6097187995910645, + "learning_rate": 0.00016320895988020218, + "loss": 0.2132, + "step": 35380 + }, + { + "epoch": 0.5520371872465215, + "grad_norm": 1.3615522384643555, + "learning_rate": 0.00016319856076204737, + "loss": 0.2751, + "step": 35390 + }, + { + "epoch": 0.5521931740188432, + "grad_norm": 0.8844773769378662, + "learning_rate": 0.00016318816164389262, + "loss": 0.22, + "step": 35400 + }, + { + "epoch": 0.5523491607911649, + "grad_norm": 0.5413331985473633, + "learning_rate": 0.0001631777625257378, + "loss": 0.2159, + "step": 35410 + }, + { + "epoch": 0.5525051475634866, + "grad_norm": 1.5456678867340088, + "learning_rate": 0.00016316736340758306, + "loss": 0.2244, + "step": 35420 + }, + { + "epoch": 0.5526611343358083, + "grad_norm": 2.1405861377716064, + "learning_rate": 0.00016315696428942825, + "loss": 0.446, + "step": 35430 + }, + { + "epoch": 0.5528171211081301, + "grad_norm": 1.4269858598709106, + "learning_rate": 0.0001631465651712735, + "loss": 0.3599, + "step": 35440 + }, + { + "epoch": 0.5529731078804517, + "grad_norm": 1.8682516813278198, + "learning_rate": 0.0001631361660531187, + "loss": 0.115, + "step": 35450 + }, + { + "epoch": 0.5531290946527735, + "grad_norm": 1.0175774097442627, + "learning_rate": 0.00016312576693496394, + "loss": 0.4003, + "step": 35460 + }, + { + "epoch": 0.5532850814250951, + "grad_norm": 2.4369170665740967, + "learning_rate": 0.00016311536781680913, + "loss": 0.2489, + "step": 35470 + }, + { + "epoch": 0.5534410681974169, + "grad_norm": 1.886022686958313, + "learning_rate": 0.00016310496869865437, + "loss": 0.4201, + "step": 35480 + }, + { + "epoch": 0.5535970549697385, + "grad_norm": 1.5811959505081177, + "learning_rate": 0.00016309456958049957, + "loss": 0.4286, + "step": 35490 + }, + { + "epoch": 0.5537530417420603, + "grad_norm": 1.971110224723816, + "learning_rate": 0.00016308417046234481, + "loss": 0.3599, + "step": 35500 + }, + { + "epoch": 0.553909028514382, + "grad_norm": 0.0335380844771862, + "learning_rate": 0.00016307377134419, + "loss": 0.3357, + "step": 35510 + }, + { + "epoch": 0.5540650152867037, + "grad_norm": 1.8519577980041504, + "learning_rate": 0.00016306337222603525, + "loss": 0.3126, + "step": 35520 + }, + { + "epoch": 0.5542210020590254, + "grad_norm": 2.1563379764556885, + "learning_rate": 0.00016305297310788045, + "loss": 0.1123, + "step": 35530 + }, + { + "epoch": 0.5543769888313471, + "grad_norm": 1.64332914352417, + "learning_rate": 0.0001630425739897257, + "loss": 0.2853, + "step": 35540 + }, + { + "epoch": 0.5545329756036688, + "grad_norm": 0.061150554567575455, + "learning_rate": 0.00016303217487157088, + "loss": 0.1943, + "step": 35550 + }, + { + "epoch": 0.5546889623759905, + "grad_norm": 1.2701060771942139, + "learning_rate": 0.00016302177575341613, + "loss": 0.2304, + "step": 35560 + }, + { + "epoch": 0.5548449491483122, + "grad_norm": 2.424860715866089, + "learning_rate": 0.00016301137663526132, + "loss": 0.2128, + "step": 35570 + }, + { + "epoch": 0.5550009359206339, + "grad_norm": 0.6803575158119202, + "learning_rate": 0.00016300097751710657, + "loss": 0.1145, + "step": 35580 + }, + { + "epoch": 0.5551569226929557, + "grad_norm": 1.2855092287063599, + "learning_rate": 0.00016299057839895176, + "loss": 0.2448, + "step": 35590 + }, + { + "epoch": 0.5553129094652773, + "grad_norm": 2.6340911388397217, + "learning_rate": 0.00016298017928079698, + "loss": 0.1856, + "step": 35600 + }, + { + "epoch": 0.5554688962375991, + "grad_norm": 1.776382327079773, + "learning_rate": 0.0001629697801626422, + "loss": 0.2422, + "step": 35610 + }, + { + "epoch": 0.5556248830099207, + "grad_norm": 3.0746347904205322, + "learning_rate": 0.00016295938104448742, + "loss": 0.1578, + "step": 35620 + }, + { + "epoch": 0.5557808697822425, + "grad_norm": 1.4887659549713135, + "learning_rate": 0.00016294898192633267, + "loss": 0.2385, + "step": 35630 + }, + { + "epoch": 0.5559368565545642, + "grad_norm": 0.5371220707893372, + "learning_rate": 0.00016293858280817786, + "loss": 0.2361, + "step": 35640 + }, + { + "epoch": 0.5560928433268859, + "grad_norm": 0.18604904413223267, + "learning_rate": 0.0001629281836900231, + "loss": 0.0828, + "step": 35650 + }, + { + "epoch": 0.5562488300992076, + "grad_norm": 2.267854690551758, + "learning_rate": 0.0001629177845718683, + "loss": 0.413, + "step": 35660 + }, + { + "epoch": 0.5564048168715293, + "grad_norm": 1.9755452871322632, + "learning_rate": 0.00016290738545371355, + "loss": 0.146, + "step": 35670 + }, + { + "epoch": 0.556560803643851, + "grad_norm": 2.430293321609497, + "learning_rate": 0.00016289698633555874, + "loss": 0.2677, + "step": 35680 + }, + { + "epoch": 0.5567167904161727, + "grad_norm": 0.6319543719291687, + "learning_rate": 0.000162886587217404, + "loss": 0.1481, + "step": 35690 + }, + { + "epoch": 0.5568727771884944, + "grad_norm": 0.17439277470111847, + "learning_rate": 0.00016287618809924918, + "loss": 0.1641, + "step": 35700 + }, + { + "epoch": 0.5570287639608161, + "grad_norm": 0.350175142288208, + "learning_rate": 0.00016286578898109443, + "loss": 0.393, + "step": 35710 + }, + { + "epoch": 0.5571847507331378, + "grad_norm": 2.641941547393799, + "learning_rate": 0.00016285538986293962, + "loss": 0.1907, + "step": 35720 + }, + { + "epoch": 0.5573407375054595, + "grad_norm": 0.1910303384065628, + "learning_rate": 0.00016284499074478487, + "loss": 0.2248, + "step": 35730 + }, + { + "epoch": 0.5574967242777813, + "grad_norm": 1.0180896520614624, + "learning_rate": 0.00016283459162663006, + "loss": 0.1924, + "step": 35740 + }, + { + "epoch": 0.557652711050103, + "grad_norm": 1.2047260999679565, + "learning_rate": 0.0001628241925084753, + "loss": 0.2031, + "step": 35750 + }, + { + "epoch": 0.5578086978224247, + "grad_norm": 1.8182405233383179, + "learning_rate": 0.0001628137933903205, + "loss": 0.1573, + "step": 35760 + }, + { + "epoch": 0.5579646845947464, + "grad_norm": 0.5485963821411133, + "learning_rate": 0.00016280339427216574, + "loss": 0.152, + "step": 35770 + }, + { + "epoch": 0.5581206713670681, + "grad_norm": 0.1674145758152008, + "learning_rate": 0.00016279299515401094, + "loss": 0.2388, + "step": 35780 + }, + { + "epoch": 0.5582766581393898, + "grad_norm": 0.36187657713890076, + "learning_rate": 0.00016278259603585618, + "loss": 0.1618, + "step": 35790 + }, + { + "epoch": 0.5584326449117115, + "grad_norm": 0.013768521137535572, + "learning_rate": 0.00016277219691770138, + "loss": 0.3666, + "step": 35800 + }, + { + "epoch": 0.5585886316840332, + "grad_norm": 2.8602802753448486, + "learning_rate": 0.00016276179779954662, + "loss": 0.2065, + "step": 35810 + }, + { + "epoch": 0.5587446184563549, + "grad_norm": 1.7925455570220947, + "learning_rate": 0.00016275139868139182, + "loss": 0.1729, + "step": 35820 + }, + { + "epoch": 0.5589006052286766, + "grad_norm": 2.910456418991089, + "learning_rate": 0.00016274099956323706, + "loss": 0.3734, + "step": 35830 + }, + { + "epoch": 0.5590565920009983, + "grad_norm": 0.8378308415412903, + "learning_rate": 0.00016273060044508225, + "loss": 0.2255, + "step": 35840 + }, + { + "epoch": 0.55921257877332, + "grad_norm": 0.409534752368927, + "learning_rate": 0.0001627202013269275, + "loss": 0.2046, + "step": 35850 + }, + { + "epoch": 0.5593685655456417, + "grad_norm": 0.49498099088668823, + "learning_rate": 0.0001627098022087727, + "loss": 0.1058, + "step": 35860 + }, + { + "epoch": 0.5595245523179634, + "grad_norm": 1.1761783361434937, + "learning_rate": 0.00016269940309061794, + "loss": 0.1566, + "step": 35870 + }, + { + "epoch": 0.5596805390902851, + "grad_norm": 3.8204751014709473, + "learning_rate": 0.00016268900397246313, + "loss": 0.2646, + "step": 35880 + }, + { + "epoch": 0.5598365258626069, + "grad_norm": 0.9882522225379944, + "learning_rate": 0.00016267860485430838, + "loss": 0.1756, + "step": 35890 + }, + { + "epoch": 0.5599925126349286, + "grad_norm": 1.1832259893417358, + "learning_rate": 0.00016266820573615357, + "loss": 0.1385, + "step": 35900 + }, + { + "epoch": 0.5601484994072503, + "grad_norm": 0.7638296484947205, + "learning_rate": 0.00016265780661799882, + "loss": 0.5679, + "step": 35910 + }, + { + "epoch": 0.560304486179572, + "grad_norm": 0.8551504611968994, + "learning_rate": 0.000162647407499844, + "loss": 0.2981, + "step": 35920 + }, + { + "epoch": 0.5604604729518937, + "grad_norm": 5.499948501586914, + "learning_rate": 0.00016263700838168926, + "loss": 0.1429, + "step": 35930 + }, + { + "epoch": 0.5606164597242154, + "grad_norm": 3.0494496822357178, + "learning_rate": 0.00016262660926353445, + "loss": 0.1322, + "step": 35940 + }, + { + "epoch": 0.5607724464965371, + "grad_norm": 0.2482384592294693, + "learning_rate": 0.0001626162101453797, + "loss": 0.2473, + "step": 35950 + }, + { + "epoch": 0.5609284332688588, + "grad_norm": 0.009750776924192905, + "learning_rate": 0.0001626058110272249, + "loss": 0.0429, + "step": 35960 + }, + { + "epoch": 0.5610844200411805, + "grad_norm": 0.5742604732513428, + "learning_rate": 0.00016259541190907014, + "loss": 0.4565, + "step": 35970 + }, + { + "epoch": 0.5612404068135022, + "grad_norm": 0.7810243368148804, + "learning_rate": 0.00016258501279091533, + "loss": 0.2107, + "step": 35980 + }, + { + "epoch": 0.5613963935858239, + "grad_norm": 2.785747528076172, + "learning_rate": 0.00016257461367276058, + "loss": 0.5087, + "step": 35990 + }, + { + "epoch": 0.5615523803581456, + "grad_norm": 0.8806902766227722, + "learning_rate": 0.00016256421455460577, + "loss": 0.2937, + "step": 36000 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.192248406016e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8df518e41a7d7fbc07133dd111b5afa296e4e151 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b9a92b2a298753078ccae56831579c2802fc8daa98e504eb13e1aff01569c127 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..f630315cc451ad653be9fda3468abb18bcbe6f8f --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f3227a142f8b34ffe6c1dfa1b60c90c1f762c58da986fe563d54b8f58afd70d +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..cb7397b171a14ad9103dc3eaf1ec6aadff7f3949 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..183420c97bb0e0af9ff72f5716b53d577dfc7cf8 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..024875ba7c7caea773293e256bf7c0aa3d497919 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/trainer_state.json @@ -0,0 +1,25584 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.569351718974231, + "eval_steps": 500, + "global_step": 36500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + }, + { + "epoch": 0.5071129968178698, + "grad_norm": 1.0075199604034424, + "learning_rate": 0.00016619350679062414, + "loss": 0.2569, + "step": 32510 + }, + { + "epoch": 0.5072689835901916, + "grad_norm": 1.3489066362380981, + "learning_rate": 0.0001661831076724694, + "loss": 0.2565, + "step": 32520 + }, + { + "epoch": 0.5074249703625132, + "grad_norm": 0.2534058094024658, + "learning_rate": 0.00016617270855431458, + "loss": 0.2292, + "step": 32530 + }, + { + "epoch": 0.507580957134835, + "grad_norm": 0.31723281741142273, + "learning_rate": 0.00016616230943615983, + "loss": 0.1467, + "step": 32540 + }, + { + "epoch": 0.5077369439071566, + "grad_norm": 0.29659849405288696, + "learning_rate": 0.00016615191031800502, + "loss": 0.2768, + "step": 32550 + }, + { + "epoch": 0.5078929306794784, + "grad_norm": 1.5246435403823853, + "learning_rate": 0.00016614151119985027, + "loss": 0.216, + "step": 32560 + }, + { + "epoch": 0.5080489174518001, + "grad_norm": 0.6026854515075684, + "learning_rate": 0.00016613111208169546, + "loss": 0.2277, + "step": 32570 + }, + { + "epoch": 0.5082049042241218, + "grad_norm": 0.41994890570640564, + "learning_rate": 0.0001661207129635407, + "loss": 0.2433, + "step": 32580 + }, + { + "epoch": 0.5083608909964435, + "grad_norm": 2.084599733352661, + "learning_rate": 0.00016611031384538593, + "loss": 0.4081, + "step": 32590 + }, + { + "epoch": 0.5085168777687652, + "grad_norm": 0.2945229709148407, + "learning_rate": 0.00016609991472723115, + "loss": 0.1329, + "step": 32600 + }, + { + "epoch": 0.508672864541087, + "grad_norm": 1.4981423616409302, + "learning_rate": 0.00016608951560907636, + "loss": 0.2855, + "step": 32610 + }, + { + "epoch": 0.5088288513134086, + "grad_norm": 0.961737871170044, + "learning_rate": 0.00016607911649092158, + "loss": 0.1654, + "step": 32620 + }, + { + "epoch": 0.5089848380857304, + "grad_norm": 1.49734628200531, + "learning_rate": 0.0001660687173727668, + "loss": 0.2081, + "step": 32630 + }, + { + "epoch": 0.509140824858052, + "grad_norm": 2.538135051727295, + "learning_rate": 0.00016605831825461202, + "loss": 0.3023, + "step": 32640 + }, + { + "epoch": 0.5092968116303738, + "grad_norm": 0.2981272041797638, + "learning_rate": 0.00016604791913645724, + "loss": 0.214, + "step": 32650 + }, + { + "epoch": 0.5094527984026954, + "grad_norm": 1.8863650560379028, + "learning_rate": 0.00016603752001830246, + "loss": 0.3363, + "step": 32660 + }, + { + "epoch": 0.5096087851750172, + "grad_norm": 0.022720983251929283, + "learning_rate": 0.00016602712090014768, + "loss": 0.2463, + "step": 32670 + }, + { + "epoch": 0.5097647719473388, + "grad_norm": 1.154977798461914, + "learning_rate": 0.0001660167217819929, + "loss": 0.4484, + "step": 32680 + }, + { + "epoch": 0.5099207587196606, + "grad_norm": 0.9470499157905579, + "learning_rate": 0.00016600632266383812, + "loss": 0.1196, + "step": 32690 + }, + { + "epoch": 0.5100767454919822, + "grad_norm": 0.19955401122570038, + "learning_rate": 0.00016599592354568334, + "loss": 0.1762, + "step": 32700 + }, + { + "epoch": 0.510232732264304, + "grad_norm": 0.4098282754421234, + "learning_rate": 0.00016598552442752856, + "loss": 0.2739, + "step": 32710 + }, + { + "epoch": 0.5103887190366257, + "grad_norm": 0.21697258949279785, + "learning_rate": 0.00016597512530937378, + "loss": 0.2555, + "step": 32720 + }, + { + "epoch": 0.5105447058089474, + "grad_norm": 4.894479751586914, + "learning_rate": 0.000165964726191219, + "loss": 0.3137, + "step": 32730 + }, + { + "epoch": 0.5107006925812692, + "grad_norm": 0.8396711349487305, + "learning_rate": 0.00016595432707306422, + "loss": 0.3105, + "step": 32740 + }, + { + "epoch": 0.5108566793535908, + "grad_norm": 1.5494674444198608, + "learning_rate": 0.00016594392795490944, + "loss": 0.2271, + "step": 32750 + }, + { + "epoch": 0.5110126661259126, + "grad_norm": 1.3409888744354248, + "learning_rate": 0.00016593352883675466, + "loss": 0.1796, + "step": 32760 + }, + { + "epoch": 0.5111686528982342, + "grad_norm": 1.7590911388397217, + "learning_rate": 0.00016592312971859988, + "loss": 0.2185, + "step": 32770 + }, + { + "epoch": 0.511324639670556, + "grad_norm": 0.7350002527236938, + "learning_rate": 0.0001659127306004451, + "loss": 0.1911, + "step": 32780 + }, + { + "epoch": 0.5114806264428776, + "grad_norm": 0.7176411747932434, + "learning_rate": 0.00016590233148229032, + "loss": 0.6918, + "step": 32790 + }, + { + "epoch": 0.5116366132151994, + "grad_norm": 2.5005321502685547, + "learning_rate": 0.00016589193236413554, + "loss": 0.2329, + "step": 32800 + }, + { + "epoch": 0.511792599987521, + "grad_norm": 1.3916250467300415, + "learning_rate": 0.00016588153324598076, + "loss": 0.2102, + "step": 32810 + }, + { + "epoch": 0.5119485867598428, + "grad_norm": 0.721671998500824, + "learning_rate": 0.00016587113412782598, + "loss": 0.2965, + "step": 32820 + }, + { + "epoch": 0.5121045735321644, + "grad_norm": 0.6120114326477051, + "learning_rate": 0.0001658607350096712, + "loss": 0.2661, + "step": 32830 + }, + { + "epoch": 0.5122605603044862, + "grad_norm": 0.29123347997665405, + "learning_rate": 0.00016585033589151642, + "loss": 0.2455, + "step": 32840 + }, + { + "epoch": 0.5124165470768078, + "grad_norm": 2.7300868034362793, + "learning_rate": 0.00016583993677336164, + "loss": 0.2049, + "step": 32850 + }, + { + "epoch": 0.5125725338491296, + "grad_norm": 1.801928997039795, + "learning_rate": 0.00016582953765520686, + "loss": 0.1725, + "step": 32860 + }, + { + "epoch": 0.5127285206214514, + "grad_norm": 0.6406471729278564, + "learning_rate": 0.00016581913853705208, + "loss": 0.164, + "step": 32870 + }, + { + "epoch": 0.512884507393773, + "grad_norm": 0.9627325534820557, + "learning_rate": 0.0001658087394188973, + "loss": 0.1728, + "step": 32880 + }, + { + "epoch": 0.5130404941660948, + "grad_norm": 1.5446962118148804, + "learning_rate": 0.00016579834030074251, + "loss": 0.2345, + "step": 32890 + }, + { + "epoch": 0.5131964809384164, + "grad_norm": 0.28092706203460693, + "learning_rate": 0.0001657879411825877, + "loss": 0.2124, + "step": 32900 + }, + { + "epoch": 0.5133524677107382, + "grad_norm": 1.7751421928405762, + "learning_rate": 0.00016577754206443295, + "loss": 0.2367, + "step": 32910 + }, + { + "epoch": 0.5135084544830598, + "grad_norm": 1.6964634656906128, + "learning_rate": 0.00016576714294627815, + "loss": 0.2361, + "step": 32920 + }, + { + "epoch": 0.5136644412553816, + "grad_norm": 0.06759823113679886, + "learning_rate": 0.0001657567438281234, + "loss": 0.2329, + "step": 32930 + }, + { + "epoch": 0.5138204280277032, + "grad_norm": 2.2519280910491943, + "learning_rate": 0.00016574634470996859, + "loss": 0.1079, + "step": 32940 + }, + { + "epoch": 0.513976414800025, + "grad_norm": 3.167576551437378, + "learning_rate": 0.00016573594559181383, + "loss": 0.1764, + "step": 32950 + }, + { + "epoch": 0.5141324015723466, + "grad_norm": 1.6734397411346436, + "learning_rate": 0.00016572554647365903, + "loss": 0.2878, + "step": 32960 + }, + { + "epoch": 0.5142883883446684, + "grad_norm": 0.35627374053001404, + "learning_rate": 0.00016571514735550427, + "loss": 0.2337, + "step": 32970 + }, + { + "epoch": 0.51444437511699, + "grad_norm": 1.6321039199829102, + "learning_rate": 0.00016570474823734946, + "loss": 0.2551, + "step": 32980 + }, + { + "epoch": 0.5146003618893118, + "grad_norm": 1.352982997894287, + "learning_rate": 0.0001656943491191947, + "loss": 0.4467, + "step": 32990 + }, + { + "epoch": 0.5147563486616334, + "grad_norm": 0.3558129072189331, + "learning_rate": 0.0001656839500010399, + "loss": 0.1913, + "step": 33000 + }, + { + "epoch": 0.5149123354339552, + "grad_norm": 0.041972286999225616, + "learning_rate": 0.00016567355088288515, + "loss": 0.1704, + "step": 33010 + }, + { + "epoch": 0.515068322206277, + "grad_norm": 1.4168939590454102, + "learning_rate": 0.00016566315176473034, + "loss": 0.4866, + "step": 33020 + }, + { + "epoch": 0.5152243089785986, + "grad_norm": 3.0042929649353027, + "learning_rate": 0.0001656527526465756, + "loss": 0.2473, + "step": 33030 + }, + { + "epoch": 0.5153802957509204, + "grad_norm": 1.027274489402771, + "learning_rate": 0.00016564235352842078, + "loss": 0.436, + "step": 33040 + }, + { + "epoch": 0.515536282523242, + "grad_norm": 1.2330195903778076, + "learning_rate": 0.00016563195441026603, + "loss": 0.4615, + "step": 33050 + }, + { + "epoch": 0.5156922692955638, + "grad_norm": 2.7145066261291504, + "learning_rate": 0.00016562155529211122, + "loss": 0.278, + "step": 33060 + }, + { + "epoch": 0.5158482560678854, + "grad_norm": 1.1273893117904663, + "learning_rate": 0.00016561115617395647, + "loss": 0.3767, + "step": 33070 + }, + { + "epoch": 0.5160042428402072, + "grad_norm": 2.466265916824341, + "learning_rate": 0.00016560075705580166, + "loss": 0.1879, + "step": 33080 + }, + { + "epoch": 0.5161602296125288, + "grad_norm": 0.3759515583515167, + "learning_rate": 0.0001655903579376469, + "loss": 0.4004, + "step": 33090 + }, + { + "epoch": 0.5163162163848506, + "grad_norm": 1.8821336030960083, + "learning_rate": 0.0001655799588194921, + "loss": 0.1622, + "step": 33100 + }, + { + "epoch": 0.5164722031571722, + "grad_norm": 0.8162096738815308, + "learning_rate": 0.00016556955970133735, + "loss": 0.1341, + "step": 33110 + }, + { + "epoch": 0.516628189929494, + "grad_norm": 0.008498801849782467, + "learning_rate": 0.00016555916058318254, + "loss": 0.2546, + "step": 33120 + }, + { + "epoch": 0.5167841767018156, + "grad_norm": 0.17536675930023193, + "learning_rate": 0.00016554876146502779, + "loss": 0.2934, + "step": 33130 + }, + { + "epoch": 0.5169401634741374, + "grad_norm": 3.539898633956909, + "learning_rate": 0.00016553836234687298, + "loss": 0.4959, + "step": 33140 + }, + { + "epoch": 0.517096150246459, + "grad_norm": 2.190887212753296, + "learning_rate": 0.00016552796322871823, + "loss": 0.2914, + "step": 33150 + }, + { + "epoch": 0.5172521370187808, + "grad_norm": 1.6388943195343018, + "learning_rate": 0.00016551756411056342, + "loss": 0.1317, + "step": 33160 + }, + { + "epoch": 0.5174081237911026, + "grad_norm": 0.33063387870788574, + "learning_rate": 0.00016550716499240866, + "loss": 0.4258, + "step": 33170 + }, + { + "epoch": 0.5175641105634242, + "grad_norm": 0.5168288946151733, + "learning_rate": 0.00016549676587425386, + "loss": 0.2548, + "step": 33180 + }, + { + "epoch": 0.517720097335746, + "grad_norm": 3.9570281505584717, + "learning_rate": 0.0001654863667560991, + "loss": 0.3179, + "step": 33190 + }, + { + "epoch": 0.5178760841080676, + "grad_norm": 1.127255916595459, + "learning_rate": 0.0001654759676379443, + "loss": 0.233, + "step": 33200 + }, + { + "epoch": 0.5180320708803894, + "grad_norm": 0.36299410462379456, + "learning_rate": 0.00016546556851978954, + "loss": 0.4614, + "step": 33210 + }, + { + "epoch": 0.518188057652711, + "grad_norm": 2.914668321609497, + "learning_rate": 0.00016545516940163474, + "loss": 0.1603, + "step": 33220 + }, + { + "epoch": 0.5183440444250328, + "grad_norm": 6.991094589233398, + "learning_rate": 0.00016544477028347998, + "loss": 0.2259, + "step": 33230 + }, + { + "epoch": 0.5185000311973544, + "grad_norm": 2.0594241619110107, + "learning_rate": 0.00016543437116532517, + "loss": 0.1885, + "step": 33240 + }, + { + "epoch": 0.5186560179696762, + "grad_norm": 0.5594693422317505, + "learning_rate": 0.00016542397204717042, + "loss": 0.4333, + "step": 33250 + }, + { + "epoch": 0.5188120047419978, + "grad_norm": 4.314854621887207, + "learning_rate": 0.00016541357292901561, + "loss": 0.2103, + "step": 33260 + }, + { + "epoch": 0.5189679915143196, + "grad_norm": 0.5609000325202942, + "learning_rate": 0.00016540317381086086, + "loss": 0.2498, + "step": 33270 + }, + { + "epoch": 0.5191239782866413, + "grad_norm": 3.680088758468628, + "learning_rate": 0.00016539277469270605, + "loss": 0.3465, + "step": 33280 + }, + { + "epoch": 0.519279965058963, + "grad_norm": 1.4993672370910645, + "learning_rate": 0.0001653823755745513, + "loss": 0.3272, + "step": 33290 + }, + { + "epoch": 0.5194359518312847, + "grad_norm": 0.18005084991455078, + "learning_rate": 0.0001653719764563965, + "loss": 0.1706, + "step": 33300 + }, + { + "epoch": 0.5195919386036064, + "grad_norm": 0.5105524659156799, + "learning_rate": 0.00016536157733824174, + "loss": 0.0937, + "step": 33310 + }, + { + "epoch": 0.5197479253759281, + "grad_norm": 1.2806589603424072, + "learning_rate": 0.00016535117822008693, + "loss": 0.1578, + "step": 33320 + }, + { + "epoch": 0.5199039121482498, + "grad_norm": 0.22015990316867828, + "learning_rate": 0.00016534077910193218, + "loss": 0.2081, + "step": 33330 + }, + { + "epoch": 0.5200598989205716, + "grad_norm": 0.34638991951942444, + "learning_rate": 0.00016533037998377737, + "loss": 0.2389, + "step": 33340 + }, + { + "epoch": 0.5202158856928932, + "grad_norm": 0.8709302544593811, + "learning_rate": 0.0001653199808656226, + "loss": 0.2112, + "step": 33350 + }, + { + "epoch": 0.520371872465215, + "grad_norm": 1.2309521436691284, + "learning_rate": 0.0001653095817474678, + "loss": 0.3448, + "step": 33360 + }, + { + "epoch": 0.5205278592375366, + "grad_norm": 1.3155115842819214, + "learning_rate": 0.00016529918262931303, + "loss": 0.1572, + "step": 33370 + }, + { + "epoch": 0.5206838460098584, + "grad_norm": 0.358751118183136, + "learning_rate": 0.00016528878351115825, + "loss": 0.3011, + "step": 33380 + }, + { + "epoch": 0.52083983278218, + "grad_norm": 3.800628662109375, + "learning_rate": 0.00016527838439300347, + "loss": 0.458, + "step": 33390 + }, + { + "epoch": 0.5209958195545018, + "grad_norm": 1.1253124475479126, + "learning_rate": 0.0001652679852748487, + "loss": 0.1238, + "step": 33400 + }, + { + "epoch": 0.5211518063268235, + "grad_norm": 0.46979108452796936, + "learning_rate": 0.0001652575861566939, + "loss": 0.0745, + "step": 33410 + }, + { + "epoch": 0.5213077930991452, + "grad_norm": 7.921492576599121, + "learning_rate": 0.00016524718703853913, + "loss": 0.4004, + "step": 33420 + }, + { + "epoch": 0.5214637798714669, + "grad_norm": 1.643127679824829, + "learning_rate": 0.00016523678792038435, + "loss": 0.2858, + "step": 33430 + }, + { + "epoch": 0.5216197666437886, + "grad_norm": 7.096631050109863, + "learning_rate": 0.0001652263888022296, + "loss": 0.4735, + "step": 33440 + }, + { + "epoch": 0.5217757534161103, + "grad_norm": 0.7280219197273254, + "learning_rate": 0.0001652159896840748, + "loss": 0.1841, + "step": 33450 + }, + { + "epoch": 0.521931740188432, + "grad_norm": 0.9415522813796997, + "learning_rate": 0.00016520559056592003, + "loss": 0.2169, + "step": 33460 + }, + { + "epoch": 0.5220877269607537, + "grad_norm": 2.264862060546875, + "learning_rate": 0.00016519519144776523, + "loss": 0.1902, + "step": 33470 + }, + { + "epoch": 0.5222437137330754, + "grad_norm": 2.4566822052001953, + "learning_rate": 0.00016518479232961047, + "loss": 0.3329, + "step": 33480 + }, + { + "epoch": 0.5223997005053972, + "grad_norm": 0.7330378890037537, + "learning_rate": 0.00016517439321145567, + "loss": 0.2341, + "step": 33490 + }, + { + "epoch": 0.5225556872777188, + "grad_norm": 1.6409542560577393, + "learning_rate": 0.0001651639940933009, + "loss": 0.3077, + "step": 33500 + }, + { + "epoch": 0.5227116740500406, + "grad_norm": 1.0647432804107666, + "learning_rate": 0.0001651535949751461, + "loss": 0.1935, + "step": 33510 + }, + { + "epoch": 0.5228676608223622, + "grad_norm": 0.507420539855957, + "learning_rate": 0.00016514319585699135, + "loss": 0.1314, + "step": 33520 + }, + { + "epoch": 0.523023647594684, + "grad_norm": 1.0064164400100708, + "learning_rate": 0.00016513279673883654, + "loss": 0.344, + "step": 33530 + }, + { + "epoch": 0.5231796343670057, + "grad_norm": 1.1936209201812744, + "learning_rate": 0.0001651223976206818, + "loss": 0.1948, + "step": 33540 + }, + { + "epoch": 0.5233356211393274, + "grad_norm": 0.04992926865816116, + "learning_rate": 0.00016511199850252698, + "loss": 0.1363, + "step": 33550 + }, + { + "epoch": 0.5234916079116491, + "grad_norm": 1.8256275653839111, + "learning_rate": 0.00016510159938437223, + "loss": 0.443, + "step": 33560 + }, + { + "epoch": 0.5236475946839708, + "grad_norm": 4.914908409118652, + "learning_rate": 0.00016509120026621742, + "loss": 0.4654, + "step": 33570 + }, + { + "epoch": 0.5238035814562925, + "grad_norm": 1.1625486612319946, + "learning_rate": 0.00016508080114806267, + "loss": 0.1974, + "step": 33580 + }, + { + "epoch": 0.5239595682286142, + "grad_norm": 6.6192522048950195, + "learning_rate": 0.00016507040202990786, + "loss": 0.2131, + "step": 33590 + }, + { + "epoch": 0.5241155550009359, + "grad_norm": 0.7369065880775452, + "learning_rate": 0.0001650600029117531, + "loss": 0.2371, + "step": 33600 + }, + { + "epoch": 0.5242715417732576, + "grad_norm": 1.5238152742385864, + "learning_rate": 0.0001650496037935983, + "loss": 0.1557, + "step": 33610 + }, + { + "epoch": 0.5244275285455793, + "grad_norm": 1.0418007373809814, + "learning_rate": 0.00016503920467544355, + "loss": 0.1878, + "step": 33620 + }, + { + "epoch": 0.524583515317901, + "grad_norm": 0.790117084980011, + "learning_rate": 0.00016502880555728874, + "loss": 0.2195, + "step": 33630 + }, + { + "epoch": 0.5247395020902228, + "grad_norm": 1.6712257862091064, + "learning_rate": 0.000165018406439134, + "loss": 0.1602, + "step": 33640 + }, + { + "epoch": 0.5248954888625444, + "grad_norm": 0.19236230850219727, + "learning_rate": 0.00016500800732097918, + "loss": 0.2526, + "step": 33650 + }, + { + "epoch": 0.5250514756348662, + "grad_norm": 1.3519701957702637, + "learning_rate": 0.00016499760820282443, + "loss": 0.4686, + "step": 33660 + }, + { + "epoch": 0.5252074624071879, + "grad_norm": 1.694342851638794, + "learning_rate": 0.00016498720908466962, + "loss": 0.1859, + "step": 33670 + }, + { + "epoch": 0.5253634491795096, + "grad_norm": 5.225239276885986, + "learning_rate": 0.00016497680996651487, + "loss": 0.2944, + "step": 33680 + }, + { + "epoch": 0.5255194359518313, + "grad_norm": 2.0208842754364014, + "learning_rate": 0.00016496641084836006, + "loss": 0.2421, + "step": 33690 + }, + { + "epoch": 0.525675422724153, + "grad_norm": 0.7954996228218079, + "learning_rate": 0.0001649560117302053, + "loss": 0.2853, + "step": 33700 + }, + { + "epoch": 0.5258314094964747, + "grad_norm": 2.296086072921753, + "learning_rate": 0.0001649456126120505, + "loss": 0.1333, + "step": 33710 + }, + { + "epoch": 0.5259873962687964, + "grad_norm": 1.1779128313064575, + "learning_rate": 0.00016493521349389574, + "loss": 0.1066, + "step": 33720 + }, + { + "epoch": 0.5261433830411181, + "grad_norm": 0.1756065934896469, + "learning_rate": 0.00016492481437574094, + "loss": 0.1352, + "step": 33730 + }, + { + "epoch": 0.5262993698134398, + "grad_norm": 0.13100725412368774, + "learning_rate": 0.00016491441525758618, + "loss": 0.2399, + "step": 33740 + }, + { + "epoch": 0.5264553565857615, + "grad_norm": 5.532008171081543, + "learning_rate": 0.00016490401613943138, + "loss": 0.2896, + "step": 33750 + }, + { + "epoch": 0.5266113433580832, + "grad_norm": 1.319886565208435, + "learning_rate": 0.00016489361702127662, + "loss": 0.3275, + "step": 33760 + }, + { + "epoch": 0.5267673301304049, + "grad_norm": 1.5550974607467651, + "learning_rate": 0.00016488321790312182, + "loss": 0.2677, + "step": 33770 + }, + { + "epoch": 0.5269233169027266, + "grad_norm": 1.8936737775802612, + "learning_rate": 0.00016487281878496706, + "loss": 0.1955, + "step": 33780 + }, + { + "epoch": 0.5270793036750484, + "grad_norm": 0.3653401732444763, + "learning_rate": 0.00016486241966681226, + "loss": 0.0723, + "step": 33790 + }, + { + "epoch": 0.52723529044737, + "grad_norm": 2.861341714859009, + "learning_rate": 0.00016485202054865747, + "loss": 0.2412, + "step": 33800 + }, + { + "epoch": 0.5273912772196918, + "grad_norm": 1.5291428565979004, + "learning_rate": 0.0001648416214305027, + "loss": 0.0871, + "step": 33810 + }, + { + "epoch": 0.5275472639920135, + "grad_norm": 1.0372581481933594, + "learning_rate": 0.00016483122231234791, + "loss": 0.4705, + "step": 33820 + }, + { + "epoch": 0.5277032507643352, + "grad_norm": 1.1943141222000122, + "learning_rate": 0.00016482082319419313, + "loss": 0.2848, + "step": 33830 + }, + { + "epoch": 0.5278592375366569, + "grad_norm": 1.9008225202560425, + "learning_rate": 0.00016481042407603835, + "loss": 0.139, + "step": 33840 + }, + { + "epoch": 0.5280152243089786, + "grad_norm": 2.132089138031006, + "learning_rate": 0.00016480002495788357, + "loss": 0.2119, + "step": 33850 + }, + { + "epoch": 0.5281712110813003, + "grad_norm": 0.24524426460266113, + "learning_rate": 0.0001647896258397288, + "loss": 0.1643, + "step": 33860 + }, + { + "epoch": 0.528327197853622, + "grad_norm": 1.6469637155532837, + "learning_rate": 0.000164779226721574, + "loss": 0.363, + "step": 33870 + }, + { + "epoch": 0.5284831846259437, + "grad_norm": 0.8767328858375549, + "learning_rate": 0.00016476882760341923, + "loss": 0.2632, + "step": 33880 + }, + { + "epoch": 0.5286391713982654, + "grad_norm": 0.06347586214542389, + "learning_rate": 0.00016475842848526445, + "loss": 0.3204, + "step": 33890 + }, + { + "epoch": 0.5287951581705871, + "grad_norm": 0.09782540798187256, + "learning_rate": 0.00016474802936710967, + "loss": 0.2211, + "step": 33900 + }, + { + "epoch": 0.5289511449429088, + "grad_norm": 3.2998859882354736, + "learning_rate": 0.0001647376302489549, + "loss": 0.265, + "step": 33910 + }, + { + "epoch": 0.5291071317152305, + "grad_norm": 0.43594226241111755, + "learning_rate": 0.0001647272311308001, + "loss": 0.194, + "step": 33920 + }, + { + "epoch": 0.5292631184875523, + "grad_norm": 1.5166605710983276, + "learning_rate": 0.00016471683201264533, + "loss": 0.2675, + "step": 33930 + }, + { + "epoch": 0.529419105259874, + "grad_norm": 0.6056640148162842, + "learning_rate": 0.00016470643289449055, + "loss": 0.0576, + "step": 33940 + }, + { + "epoch": 0.5295750920321957, + "grad_norm": 0.25410348176956177, + "learning_rate": 0.00016469603377633577, + "loss": 0.1124, + "step": 33950 + }, + { + "epoch": 0.5297310788045174, + "grad_norm": 1.770642876625061, + "learning_rate": 0.000164685634658181, + "loss": 0.3295, + "step": 33960 + }, + { + "epoch": 0.5298870655768391, + "grad_norm": 0.0607205331325531, + "learning_rate": 0.0001646752355400262, + "loss": 0.2369, + "step": 33970 + }, + { + "epoch": 0.5300430523491608, + "grad_norm": 0.5557095408439636, + "learning_rate": 0.00016466483642187143, + "loss": 0.1681, + "step": 33980 + }, + { + "epoch": 0.5301990391214825, + "grad_norm": 0.5192957520484924, + "learning_rate": 0.00016465443730371665, + "loss": 0.2649, + "step": 33990 + }, + { + "epoch": 0.5303550258938042, + "grad_norm": 0.04804835096001625, + "learning_rate": 0.00016464403818556187, + "loss": 0.2639, + "step": 34000 + }, + { + "epoch": 0.5305110126661259, + "grad_norm": 0.02673129364848137, + "learning_rate": 0.0001646336390674071, + "loss": 0.169, + "step": 34010 + }, + { + "epoch": 0.5306669994384476, + "grad_norm": 1.0084244012832642, + "learning_rate": 0.0001646232399492523, + "loss": 0.353, + "step": 34020 + }, + { + "epoch": 0.5308229862107693, + "grad_norm": 2.2202091217041016, + "learning_rate": 0.00016461284083109753, + "loss": 0.3206, + "step": 34030 + }, + { + "epoch": 0.530978972983091, + "grad_norm": 0.5573744773864746, + "learning_rate": 0.00016460244171294275, + "loss": 0.231, + "step": 34040 + }, + { + "epoch": 0.5311349597554127, + "grad_norm": 0.39700084924697876, + "learning_rate": 0.00016459204259478797, + "loss": 0.1412, + "step": 34050 + }, + { + "epoch": 0.5312909465277345, + "grad_norm": 2.582963228225708, + "learning_rate": 0.00016458164347663319, + "loss": 0.3334, + "step": 34060 + }, + { + "epoch": 0.5314469333000561, + "grad_norm": 0.22781169414520264, + "learning_rate": 0.0001645712443584784, + "loss": 0.2767, + "step": 34070 + }, + { + "epoch": 0.5316029200723779, + "grad_norm": 1.2051042318344116, + "learning_rate": 0.00016456084524032362, + "loss": 0.2045, + "step": 34080 + }, + { + "epoch": 0.5317589068446996, + "grad_norm": 0.42760100960731506, + "learning_rate": 0.00016455044612216884, + "loss": 0.1151, + "step": 34090 + }, + { + "epoch": 0.5319148936170213, + "grad_norm": 0.14440476894378662, + "learning_rate": 0.00016454004700401406, + "loss": 0.2386, + "step": 34100 + }, + { + "epoch": 0.532070880389343, + "grad_norm": 2.2777981758117676, + "learning_rate": 0.00016452964788585928, + "loss": 0.2009, + "step": 34110 + }, + { + "epoch": 0.5322268671616647, + "grad_norm": 0.9206979274749756, + "learning_rate": 0.0001645192487677045, + "loss": 0.2745, + "step": 34120 + }, + { + "epoch": 0.5323828539339864, + "grad_norm": 1.6947574615478516, + "learning_rate": 0.00016450884964954972, + "loss": 0.2584, + "step": 34130 + }, + { + "epoch": 0.5325388407063081, + "grad_norm": 0.401444673538208, + "learning_rate": 0.00016449845053139494, + "loss": 0.2218, + "step": 34140 + }, + { + "epoch": 0.5326948274786298, + "grad_norm": 0.08261553198099136, + "learning_rate": 0.00016448805141324016, + "loss": 0.2775, + "step": 34150 + }, + { + "epoch": 0.5328508142509515, + "grad_norm": 0.1017974391579628, + "learning_rate": 0.00016447765229508538, + "loss": 0.2095, + "step": 34160 + }, + { + "epoch": 0.5330068010232732, + "grad_norm": 1.3759571313858032, + "learning_rate": 0.0001644672531769306, + "loss": 0.2643, + "step": 34170 + }, + { + "epoch": 0.5331627877955949, + "grad_norm": 1.2654389142990112, + "learning_rate": 0.00016445685405877582, + "loss": 0.2949, + "step": 34180 + }, + { + "epoch": 0.5333187745679167, + "grad_norm": 1.5481843948364258, + "learning_rate": 0.00016444645494062104, + "loss": 0.1264, + "step": 34190 + }, + { + "epoch": 0.5334747613402383, + "grad_norm": 1.8094528913497925, + "learning_rate": 0.00016443605582246626, + "loss": 0.2727, + "step": 34200 + }, + { + "epoch": 0.5336307481125601, + "grad_norm": 2.224538564682007, + "learning_rate": 0.00016442565670431148, + "loss": 0.3096, + "step": 34210 + }, + { + "epoch": 0.5337867348848817, + "grad_norm": 0.6375226974487305, + "learning_rate": 0.0001644152575861567, + "loss": 0.2251, + "step": 34220 + }, + { + "epoch": 0.5339427216572035, + "grad_norm": 3.727106809616089, + "learning_rate": 0.00016440485846800192, + "loss": 0.4374, + "step": 34230 + }, + { + "epoch": 0.5340987084295252, + "grad_norm": 0.13345426321029663, + "learning_rate": 0.00016439445934984714, + "loss": 0.2011, + "step": 34240 + }, + { + "epoch": 0.5342546952018469, + "grad_norm": 2.1658668518066406, + "learning_rate": 0.00016438406023169236, + "loss": 0.2457, + "step": 34250 + }, + { + "epoch": 0.5344106819741686, + "grad_norm": 19.238407135009766, + "learning_rate": 0.00016437366111353758, + "loss": 0.2756, + "step": 34260 + }, + { + "epoch": 0.5345666687464903, + "grad_norm": 1.0292778015136719, + "learning_rate": 0.0001643632619953828, + "loss": 0.1646, + "step": 34270 + }, + { + "epoch": 0.534722655518812, + "grad_norm": 0.9372987747192383, + "learning_rate": 0.00016435286287722802, + "loss": 0.2762, + "step": 34280 + }, + { + "epoch": 0.5348786422911337, + "grad_norm": 0.3918002247810364, + "learning_rate": 0.00016434246375907324, + "loss": 0.1406, + "step": 34290 + }, + { + "epoch": 0.5350346290634554, + "grad_norm": 1.3518732786178589, + "learning_rate": 0.00016433206464091846, + "loss": 0.2992, + "step": 34300 + }, + { + "epoch": 0.5351906158357771, + "grad_norm": 0.73117595911026, + "learning_rate": 0.00016432166552276368, + "loss": 0.1172, + "step": 34310 + }, + { + "epoch": 0.5353466026080989, + "grad_norm": 0.06655958294868469, + "learning_rate": 0.0001643112664046089, + "loss": 0.1332, + "step": 34320 + }, + { + "epoch": 0.5355025893804205, + "grad_norm": 3.2587168216705322, + "learning_rate": 0.00016430086728645412, + "loss": 0.2052, + "step": 34330 + }, + { + "epoch": 0.5356585761527423, + "grad_norm": 2.6668431758880615, + "learning_rate": 0.00016429046816829934, + "loss": 0.1295, + "step": 34340 + }, + { + "epoch": 0.5358145629250639, + "grad_norm": 1.316080927848816, + "learning_rate": 0.00016428006905014456, + "loss": 0.2072, + "step": 34350 + }, + { + "epoch": 0.5359705496973857, + "grad_norm": 5.144528388977051, + "learning_rate": 0.00016426966993198977, + "loss": 0.1934, + "step": 34360 + }, + { + "epoch": 0.5361265364697073, + "grad_norm": 1.3529599905014038, + "learning_rate": 0.000164259270813835, + "loss": 0.2077, + "step": 34370 + }, + { + "epoch": 0.5362825232420291, + "grad_norm": 0.9132925868034363, + "learning_rate": 0.00016424887169568021, + "loss": 0.2404, + "step": 34380 + }, + { + "epoch": 0.5364385100143508, + "grad_norm": 1.520033597946167, + "learning_rate": 0.00016423847257752543, + "loss": 0.1765, + "step": 34390 + }, + { + "epoch": 0.5365944967866725, + "grad_norm": 2.343975305557251, + "learning_rate": 0.00016422807345937065, + "loss": 0.1959, + "step": 34400 + }, + { + "epoch": 0.5367504835589942, + "grad_norm": 0.5855118632316589, + "learning_rate": 0.00016421767434121587, + "loss": 0.0848, + "step": 34410 + }, + { + "epoch": 0.5369064703313159, + "grad_norm": 1.9108648300170898, + "learning_rate": 0.0001642072752230611, + "loss": 0.2406, + "step": 34420 + }, + { + "epoch": 0.5370624571036376, + "grad_norm": 1.0323792695999146, + "learning_rate": 0.0001641968761049063, + "loss": 0.2298, + "step": 34430 + }, + { + "epoch": 0.5372184438759593, + "grad_norm": 0.17435620725154877, + "learning_rate": 0.00016418647698675153, + "loss": 0.1396, + "step": 34440 + }, + { + "epoch": 0.537374430648281, + "grad_norm": 0.3526577353477478, + "learning_rate": 0.00016417607786859675, + "loss": 0.172, + "step": 34450 + }, + { + "epoch": 0.5375304174206027, + "grad_norm": 2.73685884475708, + "learning_rate": 0.00016416567875044197, + "loss": 0.1167, + "step": 34460 + }, + { + "epoch": 0.5376864041929245, + "grad_norm": 2.765693426132202, + "learning_rate": 0.0001641552796322872, + "loss": 0.1356, + "step": 34470 + }, + { + "epoch": 0.5378423909652461, + "grad_norm": 0.47628021240234375, + "learning_rate": 0.0001641448805141324, + "loss": 0.3104, + "step": 34480 + }, + { + "epoch": 0.5379983777375679, + "grad_norm": 1.1307590007781982, + "learning_rate": 0.00016413448139597763, + "loss": 0.3247, + "step": 34490 + }, + { + "epoch": 0.5381543645098895, + "grad_norm": 0.07136381417512894, + "learning_rate": 0.00016412408227782285, + "loss": 0.319, + "step": 34500 + }, + { + "epoch": 0.5383103512822113, + "grad_norm": 1.7850221395492554, + "learning_rate": 0.00016411368315966807, + "loss": 0.2515, + "step": 34510 + }, + { + "epoch": 0.5384663380545329, + "grad_norm": 1.3870742321014404, + "learning_rate": 0.0001641032840415133, + "loss": 0.256, + "step": 34520 + }, + { + "epoch": 0.5386223248268547, + "grad_norm": 1.1329221725463867, + "learning_rate": 0.0001640928849233585, + "loss": 0.2255, + "step": 34530 + }, + { + "epoch": 0.5387783115991764, + "grad_norm": 0.23236137628555298, + "learning_rate": 0.00016408248580520373, + "loss": 0.4053, + "step": 34540 + }, + { + "epoch": 0.5389342983714981, + "grad_norm": 3.350924253463745, + "learning_rate": 0.00016407208668704895, + "loss": 0.2816, + "step": 34550 + }, + { + "epoch": 0.5390902851438198, + "grad_norm": 1.7459170818328857, + "learning_rate": 0.00016406168756889417, + "loss": 0.0834, + "step": 34560 + }, + { + "epoch": 0.5392462719161415, + "grad_norm": 0.89351487159729, + "learning_rate": 0.0001640512884507394, + "loss": 0.1146, + "step": 34570 + }, + { + "epoch": 0.5394022586884633, + "grad_norm": 0.49410831928253174, + "learning_rate": 0.0001640408893325846, + "loss": 0.2319, + "step": 34580 + }, + { + "epoch": 0.5395582454607849, + "grad_norm": 0.009971237741410732, + "learning_rate": 0.00016403049021442983, + "loss": 0.1674, + "step": 34590 + }, + { + "epoch": 0.5397142322331067, + "grad_norm": 1.420511245727539, + "learning_rate": 0.00016402009109627505, + "loss": 0.1469, + "step": 34600 + }, + { + "epoch": 0.5398702190054283, + "grad_norm": 2.5686769485473633, + "learning_rate": 0.00016400969197812027, + "loss": 0.2607, + "step": 34610 + }, + { + "epoch": 0.5400262057777501, + "grad_norm": 2.5808913707733154, + "learning_rate": 0.00016399929285996549, + "loss": 0.4412, + "step": 34620 + }, + { + "epoch": 0.5401821925500717, + "grad_norm": 0.2778591811656952, + "learning_rate": 0.0001639888937418107, + "loss": 0.1563, + "step": 34630 + }, + { + "epoch": 0.5403381793223935, + "grad_norm": 1.6448099613189697, + "learning_rate": 0.00016397849462365592, + "loss": 0.2484, + "step": 34640 + }, + { + "epoch": 0.5404941660947151, + "grad_norm": 0.003213417250663042, + "learning_rate": 0.00016396809550550114, + "loss": 0.1582, + "step": 34650 + }, + { + "epoch": 0.5406501528670369, + "grad_norm": 2.3433990478515625, + "learning_rate": 0.00016395769638734636, + "loss": 0.2774, + "step": 34660 + }, + { + "epoch": 0.5408061396393585, + "grad_norm": 0.09544426947832108, + "learning_rate": 0.00016394729726919158, + "loss": 0.1242, + "step": 34670 + }, + { + "epoch": 0.5409621264116803, + "grad_norm": 1.8934235572814941, + "learning_rate": 0.0001639368981510368, + "loss": 0.1636, + "step": 34680 + }, + { + "epoch": 0.541118113184002, + "grad_norm": 2.8528554439544678, + "learning_rate": 0.00016392649903288202, + "loss": 0.2524, + "step": 34690 + }, + { + "epoch": 0.5412740999563237, + "grad_norm": 2.837836980819702, + "learning_rate": 0.00016391609991472722, + "loss": 0.1808, + "step": 34700 + }, + { + "epoch": 0.5414300867286455, + "grad_norm": 4.349740028381348, + "learning_rate": 0.00016390570079657246, + "loss": 0.12, + "step": 34710 + }, + { + "epoch": 0.5415860735009671, + "grad_norm": 0.6011945605278015, + "learning_rate": 0.00016389530167841765, + "loss": 0.2865, + "step": 34720 + }, + { + "epoch": 0.5417420602732889, + "grad_norm": 1.2079488039016724, + "learning_rate": 0.0001638849025602629, + "loss": 0.125, + "step": 34730 + }, + { + "epoch": 0.5418980470456105, + "grad_norm": 1.9549164772033691, + "learning_rate": 0.0001638745034421081, + "loss": 0.3285, + "step": 34740 + }, + { + "epoch": 0.5420540338179323, + "grad_norm": 0.6969407200813293, + "learning_rate": 0.00016386410432395334, + "loss": 0.3267, + "step": 34750 + }, + { + "epoch": 0.5422100205902539, + "grad_norm": 1.0561965703964233, + "learning_rate": 0.00016385370520579853, + "loss": 0.1463, + "step": 34760 + }, + { + "epoch": 0.5423660073625757, + "grad_norm": 1.8208931684494019, + "learning_rate": 0.00016384330608764378, + "loss": 0.1713, + "step": 34770 + }, + { + "epoch": 0.5425219941348973, + "grad_norm": 0.6349910497665405, + "learning_rate": 0.000163832906969489, + "loss": 0.2244, + "step": 34780 + }, + { + "epoch": 0.5426779809072191, + "grad_norm": 1.1976515054702759, + "learning_rate": 0.00016382250785133422, + "loss": 0.3407, + "step": 34790 + }, + { + "epoch": 0.5428339676795407, + "grad_norm": 0.5970319509506226, + "learning_rate": 0.00016381210873317944, + "loss": 0.1419, + "step": 34800 + }, + { + "epoch": 0.5429899544518625, + "grad_norm": 0.14623159170150757, + "learning_rate": 0.00016380170961502466, + "loss": 0.14, + "step": 34810 + }, + { + "epoch": 0.5431459412241841, + "grad_norm": 1.4332351684570312, + "learning_rate": 0.00016379131049686988, + "loss": 0.1093, + "step": 34820 + }, + { + "epoch": 0.5433019279965059, + "grad_norm": 1.1443063020706177, + "learning_rate": 0.0001637809113787151, + "loss": 0.3676, + "step": 34830 + }, + { + "epoch": 0.5434579147688277, + "grad_norm": 1.3897461891174316, + "learning_rate": 0.00016377051226056032, + "loss": 0.3713, + "step": 34840 + }, + { + "epoch": 0.5436139015411493, + "grad_norm": 1.7773199081420898, + "learning_rate": 0.00016376011314240554, + "loss": 0.2249, + "step": 34850 + }, + { + "epoch": 0.5437698883134711, + "grad_norm": 0.8524389266967773, + "learning_rate": 0.00016374971402425076, + "loss": 0.2158, + "step": 34860 + }, + { + "epoch": 0.5439258750857927, + "grad_norm": 0.5572389960289001, + "learning_rate": 0.00016373931490609598, + "loss": 0.215, + "step": 34870 + }, + { + "epoch": 0.5440818618581145, + "grad_norm": 3.0238709449768066, + "learning_rate": 0.0001637289157879412, + "loss": 0.3765, + "step": 34880 + }, + { + "epoch": 0.5442378486304361, + "grad_norm": 1.1655421257019043, + "learning_rate": 0.00016371851666978642, + "loss": 0.2491, + "step": 34890 + }, + { + "epoch": 0.5443938354027579, + "grad_norm": 0.49492090940475464, + "learning_rate": 0.00016370811755163164, + "loss": 0.1287, + "step": 34900 + }, + { + "epoch": 0.5445498221750795, + "grad_norm": 0.2732921540737152, + "learning_rate": 0.00016369771843347685, + "loss": 0.2154, + "step": 34910 + }, + { + "epoch": 0.5447058089474013, + "grad_norm": 2.5807769298553467, + "learning_rate": 0.00016368731931532207, + "loss": 0.2032, + "step": 34920 + }, + { + "epoch": 0.5448617957197229, + "grad_norm": 0.5044315457344055, + "learning_rate": 0.0001636769201971673, + "loss": 0.2595, + "step": 34930 + }, + { + "epoch": 0.5450177824920447, + "grad_norm": 2.1248972415924072, + "learning_rate": 0.00016366652107901251, + "loss": 0.5651, + "step": 34940 + }, + { + "epoch": 0.5451737692643663, + "grad_norm": 0.8391468524932861, + "learning_rate": 0.00016365612196085773, + "loss": 0.3028, + "step": 34950 + }, + { + "epoch": 0.5453297560366881, + "grad_norm": 0.36081477999687195, + "learning_rate": 0.00016364572284270295, + "loss": 0.1073, + "step": 34960 + }, + { + "epoch": 0.5454857428090097, + "grad_norm": 3.175804376602173, + "learning_rate": 0.00016363532372454817, + "loss": 0.1889, + "step": 34970 + }, + { + "epoch": 0.5456417295813315, + "grad_norm": 0.20647937059402466, + "learning_rate": 0.0001636249246063934, + "loss": 0.2416, + "step": 34980 + }, + { + "epoch": 0.5457977163536533, + "grad_norm": 2.3525943756103516, + "learning_rate": 0.0001636145254882386, + "loss": 0.3179, + "step": 34990 + }, + { + "epoch": 0.5459537031259749, + "grad_norm": 0.2641216814517975, + "learning_rate": 0.00016360412637008383, + "loss": 0.2157, + "step": 35000 + }, + { + "epoch": 0.5461096898982967, + "grad_norm": 0.09906073659658432, + "learning_rate": 0.00016359372725192905, + "loss": 0.2576, + "step": 35010 + }, + { + "epoch": 0.5462656766706183, + "grad_norm": 0.13971920311450958, + "learning_rate": 0.00016358332813377427, + "loss": 0.1298, + "step": 35020 + }, + { + "epoch": 0.5464216634429401, + "grad_norm": 1.4535638093948364, + "learning_rate": 0.0001635729290156195, + "loss": 0.2357, + "step": 35030 + }, + { + "epoch": 0.5465776502152617, + "grad_norm": 0.9107828736305237, + "learning_rate": 0.0001635625298974647, + "loss": 0.3049, + "step": 35040 + }, + { + "epoch": 0.5467336369875835, + "grad_norm": 0.0377386212348938, + "learning_rate": 0.00016355213077930993, + "loss": 0.2066, + "step": 35050 + }, + { + "epoch": 0.5468896237599051, + "grad_norm": 2.7800869941711426, + "learning_rate": 0.00016354173166115515, + "loss": 0.2974, + "step": 35060 + }, + { + "epoch": 0.5470456105322269, + "grad_norm": 1.6247998476028442, + "learning_rate": 0.00016353133254300037, + "loss": 0.3822, + "step": 35070 + }, + { + "epoch": 0.5472015973045485, + "grad_norm": 1.6479015350341797, + "learning_rate": 0.0001635209334248456, + "loss": 0.3401, + "step": 35080 + }, + { + "epoch": 0.5473575840768703, + "grad_norm": 0.1879737675189972, + "learning_rate": 0.0001635105343066908, + "loss": 0.1549, + "step": 35090 + }, + { + "epoch": 0.5475135708491919, + "grad_norm": 0.39355361461639404, + "learning_rate": 0.00016350013518853603, + "loss": 0.2032, + "step": 35100 + }, + { + "epoch": 0.5476695576215137, + "grad_norm": 2.9976983070373535, + "learning_rate": 0.00016348973607038125, + "loss": 0.3867, + "step": 35110 + }, + { + "epoch": 0.5478255443938354, + "grad_norm": 0.8242707252502441, + "learning_rate": 0.00016347933695222647, + "loss": 0.3601, + "step": 35120 + }, + { + "epoch": 0.5479815311661571, + "grad_norm": 0.874955415725708, + "learning_rate": 0.0001634689378340717, + "loss": 0.2709, + "step": 35130 + }, + { + "epoch": 0.5481375179384789, + "grad_norm": 0.05980971083045006, + "learning_rate": 0.0001634585387159169, + "loss": 0.2646, + "step": 35140 + }, + { + "epoch": 0.5482935047108005, + "grad_norm": 1.8437328338623047, + "learning_rate": 0.0001634481395977621, + "loss": 0.1965, + "step": 35150 + }, + { + "epoch": 0.5484494914831223, + "grad_norm": 0.4142405688762665, + "learning_rate": 0.00016343774047960735, + "loss": 0.1756, + "step": 35160 + }, + { + "epoch": 0.5486054782554439, + "grad_norm": 4.836324214935303, + "learning_rate": 0.00016342734136145254, + "loss": 0.2534, + "step": 35170 + }, + { + "epoch": 0.5487614650277657, + "grad_norm": 1.4933065176010132, + "learning_rate": 0.00016341694224329779, + "loss": 0.2355, + "step": 35180 + }, + { + "epoch": 0.5489174518000873, + "grad_norm": 1.9468894004821777, + "learning_rate": 0.00016340654312514298, + "loss": 0.3113, + "step": 35190 + }, + { + "epoch": 0.5490734385724091, + "grad_norm": 1.012710452079773, + "learning_rate": 0.00016339614400698822, + "loss": 0.2635, + "step": 35200 + }, + { + "epoch": 0.5492294253447307, + "grad_norm": 1.5551140308380127, + "learning_rate": 0.00016338574488883342, + "loss": 0.2951, + "step": 35210 + }, + { + "epoch": 0.5493854121170525, + "grad_norm": 1.8613696098327637, + "learning_rate": 0.00016337534577067866, + "loss": 0.2832, + "step": 35220 + }, + { + "epoch": 0.5495413988893741, + "grad_norm": 1.6748839616775513, + "learning_rate": 0.00016336494665252386, + "loss": 0.2266, + "step": 35230 + }, + { + "epoch": 0.5496973856616959, + "grad_norm": 0.4094032347202301, + "learning_rate": 0.0001633545475343691, + "loss": 0.3497, + "step": 35240 + }, + { + "epoch": 0.5498533724340176, + "grad_norm": 1.6414631605148315, + "learning_rate": 0.0001633441484162143, + "loss": 0.1763, + "step": 35250 + }, + { + "epoch": 0.5500093592063393, + "grad_norm": 1.6380645036697388, + "learning_rate": 0.00016333374929805954, + "loss": 0.1355, + "step": 35260 + }, + { + "epoch": 0.550165345978661, + "grad_norm": 0.7325630187988281, + "learning_rate": 0.00016332335017990473, + "loss": 0.3105, + "step": 35270 + }, + { + "epoch": 0.5503213327509827, + "grad_norm": 1.6015644073486328, + "learning_rate": 0.00016331295106174998, + "loss": 0.2311, + "step": 35280 + }, + { + "epoch": 0.5504773195233045, + "grad_norm": 0.9224210381507874, + "learning_rate": 0.00016330255194359517, + "loss": 0.2746, + "step": 35290 + }, + { + "epoch": 0.5506333062956261, + "grad_norm": 1.30025315284729, + "learning_rate": 0.00016329215282544042, + "loss": 0.2066, + "step": 35300 + }, + { + "epoch": 0.5507892930679479, + "grad_norm": 0.481125146150589, + "learning_rate": 0.0001632817537072856, + "loss": 0.1778, + "step": 35310 + }, + { + "epoch": 0.5509452798402695, + "grad_norm": 0.2709486782550812, + "learning_rate": 0.00016327135458913086, + "loss": 0.1131, + "step": 35320 + }, + { + "epoch": 0.5511012666125913, + "grad_norm": 0.6248563528060913, + "learning_rate": 0.00016326095547097605, + "loss": 0.4356, + "step": 35330 + }, + { + "epoch": 0.5512572533849129, + "grad_norm": 3.0947647094726562, + "learning_rate": 0.0001632505563528213, + "loss": 0.3072, + "step": 35340 + }, + { + "epoch": 0.5514132401572347, + "grad_norm": 1.009535312652588, + "learning_rate": 0.0001632401572346665, + "loss": 0.1808, + "step": 35350 + }, + { + "epoch": 0.5515692269295563, + "grad_norm": 0.042604975402355194, + "learning_rate": 0.00016322975811651174, + "loss": 0.1198, + "step": 35360 + }, + { + "epoch": 0.5517252137018781, + "grad_norm": 1.8652396202087402, + "learning_rate": 0.00016321935899835693, + "loss": 0.2176, + "step": 35370 + }, + { + "epoch": 0.5518812004741998, + "grad_norm": 4.6097187995910645, + "learning_rate": 0.00016320895988020218, + "loss": 0.2132, + "step": 35380 + }, + { + "epoch": 0.5520371872465215, + "grad_norm": 1.3615522384643555, + "learning_rate": 0.00016319856076204737, + "loss": 0.2751, + "step": 35390 + }, + { + "epoch": 0.5521931740188432, + "grad_norm": 0.8844773769378662, + "learning_rate": 0.00016318816164389262, + "loss": 0.22, + "step": 35400 + }, + { + "epoch": 0.5523491607911649, + "grad_norm": 0.5413331985473633, + "learning_rate": 0.0001631777625257378, + "loss": 0.2159, + "step": 35410 + }, + { + "epoch": 0.5525051475634866, + "grad_norm": 1.5456678867340088, + "learning_rate": 0.00016316736340758306, + "loss": 0.2244, + "step": 35420 + }, + { + "epoch": 0.5526611343358083, + "grad_norm": 2.1405861377716064, + "learning_rate": 0.00016315696428942825, + "loss": 0.446, + "step": 35430 + }, + { + "epoch": 0.5528171211081301, + "grad_norm": 1.4269858598709106, + "learning_rate": 0.0001631465651712735, + "loss": 0.3599, + "step": 35440 + }, + { + "epoch": 0.5529731078804517, + "grad_norm": 1.8682516813278198, + "learning_rate": 0.0001631361660531187, + "loss": 0.115, + "step": 35450 + }, + { + "epoch": 0.5531290946527735, + "grad_norm": 1.0175774097442627, + "learning_rate": 0.00016312576693496394, + "loss": 0.4003, + "step": 35460 + }, + { + "epoch": 0.5532850814250951, + "grad_norm": 2.4369170665740967, + "learning_rate": 0.00016311536781680913, + "loss": 0.2489, + "step": 35470 + }, + { + "epoch": 0.5534410681974169, + "grad_norm": 1.886022686958313, + "learning_rate": 0.00016310496869865437, + "loss": 0.4201, + "step": 35480 + }, + { + "epoch": 0.5535970549697385, + "grad_norm": 1.5811959505081177, + "learning_rate": 0.00016309456958049957, + "loss": 0.4286, + "step": 35490 + }, + { + "epoch": 0.5537530417420603, + "grad_norm": 1.971110224723816, + "learning_rate": 0.00016308417046234481, + "loss": 0.3599, + "step": 35500 + }, + { + "epoch": 0.553909028514382, + "grad_norm": 0.0335380844771862, + "learning_rate": 0.00016307377134419, + "loss": 0.3357, + "step": 35510 + }, + { + "epoch": 0.5540650152867037, + "grad_norm": 1.8519577980041504, + "learning_rate": 0.00016306337222603525, + "loss": 0.3126, + "step": 35520 + }, + { + "epoch": 0.5542210020590254, + "grad_norm": 2.1563379764556885, + "learning_rate": 0.00016305297310788045, + "loss": 0.1123, + "step": 35530 + }, + { + "epoch": 0.5543769888313471, + "grad_norm": 1.64332914352417, + "learning_rate": 0.0001630425739897257, + "loss": 0.2853, + "step": 35540 + }, + { + "epoch": 0.5545329756036688, + "grad_norm": 0.061150554567575455, + "learning_rate": 0.00016303217487157088, + "loss": 0.1943, + "step": 35550 + }, + { + "epoch": 0.5546889623759905, + "grad_norm": 1.2701060771942139, + "learning_rate": 0.00016302177575341613, + "loss": 0.2304, + "step": 35560 + }, + { + "epoch": 0.5548449491483122, + "grad_norm": 2.424860715866089, + "learning_rate": 0.00016301137663526132, + "loss": 0.2128, + "step": 35570 + }, + { + "epoch": 0.5550009359206339, + "grad_norm": 0.6803575158119202, + "learning_rate": 0.00016300097751710657, + "loss": 0.1145, + "step": 35580 + }, + { + "epoch": 0.5551569226929557, + "grad_norm": 1.2855092287063599, + "learning_rate": 0.00016299057839895176, + "loss": 0.2448, + "step": 35590 + }, + { + "epoch": 0.5553129094652773, + "grad_norm": 2.6340911388397217, + "learning_rate": 0.00016298017928079698, + "loss": 0.1856, + "step": 35600 + }, + { + "epoch": 0.5554688962375991, + "grad_norm": 1.776382327079773, + "learning_rate": 0.0001629697801626422, + "loss": 0.2422, + "step": 35610 + }, + { + "epoch": 0.5556248830099207, + "grad_norm": 3.0746347904205322, + "learning_rate": 0.00016295938104448742, + "loss": 0.1578, + "step": 35620 + }, + { + "epoch": 0.5557808697822425, + "grad_norm": 1.4887659549713135, + "learning_rate": 0.00016294898192633267, + "loss": 0.2385, + "step": 35630 + }, + { + "epoch": 0.5559368565545642, + "grad_norm": 0.5371220707893372, + "learning_rate": 0.00016293858280817786, + "loss": 0.2361, + "step": 35640 + }, + { + "epoch": 0.5560928433268859, + "grad_norm": 0.18604904413223267, + "learning_rate": 0.0001629281836900231, + "loss": 0.0828, + "step": 35650 + }, + { + "epoch": 0.5562488300992076, + "grad_norm": 2.267854690551758, + "learning_rate": 0.0001629177845718683, + "loss": 0.413, + "step": 35660 + }, + { + "epoch": 0.5564048168715293, + "grad_norm": 1.9755452871322632, + "learning_rate": 0.00016290738545371355, + "loss": 0.146, + "step": 35670 + }, + { + "epoch": 0.556560803643851, + "grad_norm": 2.430293321609497, + "learning_rate": 0.00016289698633555874, + "loss": 0.2677, + "step": 35680 + }, + { + "epoch": 0.5567167904161727, + "grad_norm": 0.6319543719291687, + "learning_rate": 0.000162886587217404, + "loss": 0.1481, + "step": 35690 + }, + { + "epoch": 0.5568727771884944, + "grad_norm": 0.17439277470111847, + "learning_rate": 0.00016287618809924918, + "loss": 0.1641, + "step": 35700 + }, + { + "epoch": 0.5570287639608161, + "grad_norm": 0.350175142288208, + "learning_rate": 0.00016286578898109443, + "loss": 0.393, + "step": 35710 + }, + { + "epoch": 0.5571847507331378, + "grad_norm": 2.641941547393799, + "learning_rate": 0.00016285538986293962, + "loss": 0.1907, + "step": 35720 + }, + { + "epoch": 0.5573407375054595, + "grad_norm": 0.1910303384065628, + "learning_rate": 0.00016284499074478487, + "loss": 0.2248, + "step": 35730 + }, + { + "epoch": 0.5574967242777813, + "grad_norm": 1.0180896520614624, + "learning_rate": 0.00016283459162663006, + "loss": 0.1924, + "step": 35740 + }, + { + "epoch": 0.557652711050103, + "grad_norm": 1.2047260999679565, + "learning_rate": 0.0001628241925084753, + "loss": 0.2031, + "step": 35750 + }, + { + "epoch": 0.5578086978224247, + "grad_norm": 1.8182405233383179, + "learning_rate": 0.0001628137933903205, + "loss": 0.1573, + "step": 35760 + }, + { + "epoch": 0.5579646845947464, + "grad_norm": 0.5485963821411133, + "learning_rate": 0.00016280339427216574, + "loss": 0.152, + "step": 35770 + }, + { + "epoch": 0.5581206713670681, + "grad_norm": 0.1674145758152008, + "learning_rate": 0.00016279299515401094, + "loss": 0.2388, + "step": 35780 + }, + { + "epoch": 0.5582766581393898, + "grad_norm": 0.36187657713890076, + "learning_rate": 0.00016278259603585618, + "loss": 0.1618, + "step": 35790 + }, + { + "epoch": 0.5584326449117115, + "grad_norm": 0.013768521137535572, + "learning_rate": 0.00016277219691770138, + "loss": 0.3666, + "step": 35800 + }, + { + "epoch": 0.5585886316840332, + "grad_norm": 2.8602802753448486, + "learning_rate": 0.00016276179779954662, + "loss": 0.2065, + "step": 35810 + }, + { + "epoch": 0.5587446184563549, + "grad_norm": 1.7925455570220947, + "learning_rate": 0.00016275139868139182, + "loss": 0.1729, + "step": 35820 + }, + { + "epoch": 0.5589006052286766, + "grad_norm": 2.910456418991089, + "learning_rate": 0.00016274099956323706, + "loss": 0.3734, + "step": 35830 + }, + { + "epoch": 0.5590565920009983, + "grad_norm": 0.8378308415412903, + "learning_rate": 0.00016273060044508225, + "loss": 0.2255, + "step": 35840 + }, + { + "epoch": 0.55921257877332, + "grad_norm": 0.409534752368927, + "learning_rate": 0.0001627202013269275, + "loss": 0.2046, + "step": 35850 + }, + { + "epoch": 0.5593685655456417, + "grad_norm": 0.49498099088668823, + "learning_rate": 0.0001627098022087727, + "loss": 0.1058, + "step": 35860 + }, + { + "epoch": 0.5595245523179634, + "grad_norm": 1.1761783361434937, + "learning_rate": 0.00016269940309061794, + "loss": 0.1566, + "step": 35870 + }, + { + "epoch": 0.5596805390902851, + "grad_norm": 3.8204751014709473, + "learning_rate": 0.00016268900397246313, + "loss": 0.2646, + "step": 35880 + }, + { + "epoch": 0.5598365258626069, + "grad_norm": 0.9882522225379944, + "learning_rate": 0.00016267860485430838, + "loss": 0.1756, + "step": 35890 + }, + { + "epoch": 0.5599925126349286, + "grad_norm": 1.1832259893417358, + "learning_rate": 0.00016266820573615357, + "loss": 0.1385, + "step": 35900 + }, + { + "epoch": 0.5601484994072503, + "grad_norm": 0.7638296484947205, + "learning_rate": 0.00016265780661799882, + "loss": 0.5679, + "step": 35910 + }, + { + "epoch": 0.560304486179572, + "grad_norm": 0.8551504611968994, + "learning_rate": 0.000162647407499844, + "loss": 0.2981, + "step": 35920 + }, + { + "epoch": 0.5604604729518937, + "grad_norm": 5.499948501586914, + "learning_rate": 0.00016263700838168926, + "loss": 0.1429, + "step": 35930 + }, + { + "epoch": 0.5606164597242154, + "grad_norm": 3.0494496822357178, + "learning_rate": 0.00016262660926353445, + "loss": 0.1322, + "step": 35940 + }, + { + "epoch": 0.5607724464965371, + "grad_norm": 0.2482384592294693, + "learning_rate": 0.0001626162101453797, + "loss": 0.2473, + "step": 35950 + }, + { + "epoch": 0.5609284332688588, + "grad_norm": 0.009750776924192905, + "learning_rate": 0.0001626058110272249, + "loss": 0.0429, + "step": 35960 + }, + { + "epoch": 0.5610844200411805, + "grad_norm": 0.5742604732513428, + "learning_rate": 0.00016259541190907014, + "loss": 0.4565, + "step": 35970 + }, + { + "epoch": 0.5612404068135022, + "grad_norm": 0.7810243368148804, + "learning_rate": 0.00016258501279091533, + "loss": 0.2107, + "step": 35980 + }, + { + "epoch": 0.5613963935858239, + "grad_norm": 2.785747528076172, + "learning_rate": 0.00016257461367276058, + "loss": 0.5087, + "step": 35990 + }, + { + "epoch": 0.5615523803581456, + "grad_norm": 0.8806902766227722, + "learning_rate": 0.00016256421455460577, + "loss": 0.2937, + "step": 36000 + }, + { + "epoch": 0.5617083671304673, + "grad_norm": 1.857373833656311, + "learning_rate": 0.00016255381543645102, + "loss": 0.5042, + "step": 36010 + }, + { + "epoch": 0.561864353902789, + "grad_norm": 0.5333901047706604, + "learning_rate": 0.0001625434163182962, + "loss": 0.1678, + "step": 36020 + }, + { + "epoch": 0.5620203406751108, + "grad_norm": 11.023160934448242, + "learning_rate": 0.00016253301720014145, + "loss": 0.2721, + "step": 36030 + }, + { + "epoch": 0.5621763274474325, + "grad_norm": 2.6214029788970947, + "learning_rate": 0.00016252261808198665, + "loss": 0.1675, + "step": 36040 + }, + { + "epoch": 0.5623323142197542, + "grad_norm": 0.25980064272880554, + "learning_rate": 0.0001625122189638319, + "loss": 0.1832, + "step": 36050 + }, + { + "epoch": 0.5624883009920759, + "grad_norm": 1.3559473752975464, + "learning_rate": 0.0001625018198456771, + "loss": 0.121, + "step": 36060 + }, + { + "epoch": 0.5626442877643976, + "grad_norm": 0.3558153212070465, + "learning_rate": 0.0001624914207275223, + "loss": 0.2797, + "step": 36070 + }, + { + "epoch": 0.5628002745367193, + "grad_norm": 0.6314427256584167, + "learning_rate": 0.00016248102160936753, + "loss": 0.3455, + "step": 36080 + }, + { + "epoch": 0.562956261309041, + "grad_norm": 0.16120101511478424, + "learning_rate": 0.00016247062249121275, + "loss": 0.3669, + "step": 36090 + }, + { + "epoch": 0.5631122480813627, + "grad_norm": 1.041572093963623, + "learning_rate": 0.00016246022337305797, + "loss": 0.2744, + "step": 36100 + }, + { + "epoch": 0.5632682348536844, + "grad_norm": 0.2645890414714813, + "learning_rate": 0.00016244982425490318, + "loss": 0.1487, + "step": 36110 + }, + { + "epoch": 0.5634242216260061, + "grad_norm": 5.698698043823242, + "learning_rate": 0.0001624394251367484, + "loss": 0.173, + "step": 36120 + }, + { + "epoch": 0.5635802083983278, + "grad_norm": 0.3299804627895355, + "learning_rate": 0.00016242902601859362, + "loss": 0.1965, + "step": 36130 + }, + { + "epoch": 0.5637361951706495, + "grad_norm": 1.2476481199264526, + "learning_rate": 0.00016241862690043884, + "loss": 0.4097, + "step": 36140 + }, + { + "epoch": 0.5638921819429712, + "grad_norm": 0.30926263332366943, + "learning_rate": 0.00016240822778228406, + "loss": 0.1416, + "step": 36150 + }, + { + "epoch": 0.564048168715293, + "grad_norm": 0.05872740224003792, + "learning_rate": 0.00016239782866412928, + "loss": 0.1449, + "step": 36160 + }, + { + "epoch": 0.5642041554876146, + "grad_norm": 0.2292211800813675, + "learning_rate": 0.0001623874295459745, + "loss": 0.2557, + "step": 36170 + }, + { + "epoch": 0.5643601422599364, + "grad_norm": 1.7822531461715698, + "learning_rate": 0.00016237703042781972, + "loss": 0.2953, + "step": 36180 + }, + { + "epoch": 0.5645161290322581, + "grad_norm": 2.4908461570739746, + "learning_rate": 0.00016236663130966494, + "loss": 0.3888, + "step": 36190 + }, + { + "epoch": 0.5646721158045798, + "grad_norm": 0.08363594114780426, + "learning_rate": 0.00016235623219151016, + "loss": 0.2293, + "step": 36200 + }, + { + "epoch": 0.5648281025769015, + "grad_norm": 1.4830002784729004, + "learning_rate": 0.00016234583307335538, + "loss": 0.2348, + "step": 36210 + }, + { + "epoch": 0.5649840893492232, + "grad_norm": 1.8443071842193604, + "learning_rate": 0.0001623354339552006, + "loss": 0.3392, + "step": 36220 + }, + { + "epoch": 0.5651400761215449, + "grad_norm": 2.8050875663757324, + "learning_rate": 0.00016232503483704582, + "loss": 0.2426, + "step": 36230 + }, + { + "epoch": 0.5652960628938666, + "grad_norm": 3.6627259254455566, + "learning_rate": 0.00016231463571889104, + "loss": 0.2118, + "step": 36240 + }, + { + "epoch": 0.5654520496661883, + "grad_norm": 0.3443094789981842, + "learning_rate": 0.00016230423660073626, + "loss": 0.2418, + "step": 36250 + }, + { + "epoch": 0.56560803643851, + "grad_norm": 1.7411112785339355, + "learning_rate": 0.00016229383748258148, + "loss": 0.1628, + "step": 36260 + }, + { + "epoch": 0.5657640232108317, + "grad_norm": 1.1321816444396973, + "learning_rate": 0.0001622834383644267, + "loss": 0.2732, + "step": 36270 + }, + { + "epoch": 0.5659200099831534, + "grad_norm": 2.9686615467071533, + "learning_rate": 0.00016227303924627192, + "loss": 0.3036, + "step": 36280 + }, + { + "epoch": 0.5660759967554752, + "grad_norm": 5.03530740737915, + "learning_rate": 0.00016226264012811714, + "loss": 0.2143, + "step": 36290 + }, + { + "epoch": 0.5662319835277968, + "grad_norm": 2.3943281173706055, + "learning_rate": 0.00016225224100996236, + "loss": 0.2629, + "step": 36300 + }, + { + "epoch": 0.5663879703001186, + "grad_norm": 1.812828779220581, + "learning_rate": 0.00016224184189180758, + "loss": 0.3892, + "step": 36310 + }, + { + "epoch": 0.5665439570724402, + "grad_norm": 1.4708483219146729, + "learning_rate": 0.0001622314427736528, + "loss": 0.2118, + "step": 36320 + }, + { + "epoch": 0.566699943844762, + "grad_norm": 8.99713134765625, + "learning_rate": 0.00016222104365549802, + "loss": 0.3185, + "step": 36330 + }, + { + "epoch": 0.5668559306170837, + "grad_norm": 1.7472341060638428, + "learning_rate": 0.00016221064453734324, + "loss": 0.1659, + "step": 36340 + }, + { + "epoch": 0.5670119173894054, + "grad_norm": 0.8059778213500977, + "learning_rate": 0.00016220024541918846, + "loss": 0.0621, + "step": 36350 + }, + { + "epoch": 0.5671679041617271, + "grad_norm": 4.0501861572265625, + "learning_rate": 0.00016218984630103368, + "loss": 0.1399, + "step": 36360 + }, + { + "epoch": 0.5673238909340488, + "grad_norm": 0.014400321058928967, + "learning_rate": 0.0001621794471828789, + "loss": 0.2372, + "step": 36370 + }, + { + "epoch": 0.5674798777063705, + "grad_norm": 0.4061933755874634, + "learning_rate": 0.00016216904806472412, + "loss": 0.3854, + "step": 36380 + }, + { + "epoch": 0.5676358644786922, + "grad_norm": 0.4940955340862274, + "learning_rate": 0.00016215864894656933, + "loss": 0.1689, + "step": 36390 + }, + { + "epoch": 0.567791851251014, + "grad_norm": 0.3220517635345459, + "learning_rate": 0.00016214824982841455, + "loss": 0.1912, + "step": 36400 + }, + { + "epoch": 0.5679478380233356, + "grad_norm": 2.0848920345306396, + "learning_rate": 0.00016213785071025977, + "loss": 0.1639, + "step": 36410 + }, + { + "epoch": 0.5681038247956574, + "grad_norm": 0.04129406809806824, + "learning_rate": 0.000162127451592105, + "loss": 0.1378, + "step": 36420 + }, + { + "epoch": 0.568259811567979, + "grad_norm": 1.9343401193618774, + "learning_rate": 0.0001621170524739502, + "loss": 0.3759, + "step": 36430 + }, + { + "epoch": 0.5684157983403008, + "grad_norm": 1.0625615119934082, + "learning_rate": 0.00016210665335579543, + "loss": 0.2074, + "step": 36440 + }, + { + "epoch": 0.5685717851126224, + "grad_norm": 0.14784491062164307, + "learning_rate": 0.00016209625423764065, + "loss": 0.0981, + "step": 36450 + }, + { + "epoch": 0.5687277718849442, + "grad_norm": 1.8117821216583252, + "learning_rate": 0.00016208585511948587, + "loss": 0.2755, + "step": 36460 + }, + { + "epoch": 0.5688837586572658, + "grad_norm": 0.061010006815195084, + "learning_rate": 0.0001620754560013311, + "loss": 0.1815, + "step": 36470 + }, + { + "epoch": 0.5690397454295876, + "grad_norm": 0.44453224539756775, + "learning_rate": 0.0001620650568831763, + "loss": 0.0891, + "step": 36480 + }, + { + "epoch": 0.5691957322019093, + "grad_norm": 0.6130645871162415, + "learning_rate": 0.00016205465776502153, + "loss": 0.2778, + "step": 36490 + }, + { + "epoch": 0.569351718974231, + "grad_norm": 5.408796787261963, + "learning_rate": 0.00016204425864686678, + "loss": 0.2301, + "step": 36500 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.222696300544e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-36500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ca46df3f321070424dcbb409e40fce08aa9fca00 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:790f8e1fe640adc05bd557c9a922d8731e78d1a6bfaca3556467f01445237dd7 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..7d350d8ea7335006b6f63b391f07399beb48d437 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:545cf0dd7b9ae0a2c0eee0bc951ec31f5f4445341616cb7925051a54600d13f5 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..24982340f37e4eb93c51cc35293dfe8986a6d813 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..02a3415f291e9851deb9bfdf36ddbf269c4af1d2 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..0efa3de5050bc65e266655075709e64884603987 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/trainer_state.json @@ -0,0 +1,25934 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5771510575903164, + "eval_steps": 500, + "global_step": 37000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + }, + { + "epoch": 0.5071129968178698, + "grad_norm": 1.0075199604034424, + "learning_rate": 0.00016619350679062414, + "loss": 0.2569, + "step": 32510 + }, + { + "epoch": 0.5072689835901916, + "grad_norm": 1.3489066362380981, + "learning_rate": 0.0001661831076724694, + "loss": 0.2565, + "step": 32520 + }, + { + "epoch": 0.5074249703625132, + "grad_norm": 0.2534058094024658, + "learning_rate": 0.00016617270855431458, + "loss": 0.2292, + "step": 32530 + }, + { + "epoch": 0.507580957134835, + "grad_norm": 0.31723281741142273, + "learning_rate": 0.00016616230943615983, + "loss": 0.1467, + "step": 32540 + }, + { + "epoch": 0.5077369439071566, + "grad_norm": 0.29659849405288696, + "learning_rate": 0.00016615191031800502, + "loss": 0.2768, + "step": 32550 + }, + { + "epoch": 0.5078929306794784, + "grad_norm": 1.5246435403823853, + "learning_rate": 0.00016614151119985027, + "loss": 0.216, + "step": 32560 + }, + { + "epoch": 0.5080489174518001, + "grad_norm": 0.6026854515075684, + "learning_rate": 0.00016613111208169546, + "loss": 0.2277, + "step": 32570 + }, + { + "epoch": 0.5082049042241218, + "grad_norm": 0.41994890570640564, + "learning_rate": 0.0001661207129635407, + "loss": 0.2433, + "step": 32580 + }, + { + "epoch": 0.5083608909964435, + "grad_norm": 2.084599733352661, + "learning_rate": 0.00016611031384538593, + "loss": 0.4081, + "step": 32590 + }, + { + "epoch": 0.5085168777687652, + "grad_norm": 0.2945229709148407, + "learning_rate": 0.00016609991472723115, + "loss": 0.1329, + "step": 32600 + }, + { + "epoch": 0.508672864541087, + "grad_norm": 1.4981423616409302, + "learning_rate": 0.00016608951560907636, + "loss": 0.2855, + "step": 32610 + }, + { + "epoch": 0.5088288513134086, + "grad_norm": 0.961737871170044, + "learning_rate": 0.00016607911649092158, + "loss": 0.1654, + "step": 32620 + }, + { + "epoch": 0.5089848380857304, + "grad_norm": 1.49734628200531, + "learning_rate": 0.0001660687173727668, + "loss": 0.2081, + "step": 32630 + }, + { + "epoch": 0.509140824858052, + "grad_norm": 2.538135051727295, + "learning_rate": 0.00016605831825461202, + "loss": 0.3023, + "step": 32640 + }, + { + "epoch": 0.5092968116303738, + "grad_norm": 0.2981272041797638, + "learning_rate": 0.00016604791913645724, + "loss": 0.214, + "step": 32650 + }, + { + "epoch": 0.5094527984026954, + "grad_norm": 1.8863650560379028, + "learning_rate": 0.00016603752001830246, + "loss": 0.3363, + "step": 32660 + }, + { + "epoch": 0.5096087851750172, + "grad_norm": 0.022720983251929283, + "learning_rate": 0.00016602712090014768, + "loss": 0.2463, + "step": 32670 + }, + { + "epoch": 0.5097647719473388, + "grad_norm": 1.154977798461914, + "learning_rate": 0.0001660167217819929, + "loss": 0.4484, + "step": 32680 + }, + { + "epoch": 0.5099207587196606, + "grad_norm": 0.9470499157905579, + "learning_rate": 0.00016600632266383812, + "loss": 0.1196, + "step": 32690 + }, + { + "epoch": 0.5100767454919822, + "grad_norm": 0.19955401122570038, + "learning_rate": 0.00016599592354568334, + "loss": 0.1762, + "step": 32700 + }, + { + "epoch": 0.510232732264304, + "grad_norm": 0.4098282754421234, + "learning_rate": 0.00016598552442752856, + "loss": 0.2739, + "step": 32710 + }, + { + "epoch": 0.5103887190366257, + "grad_norm": 0.21697258949279785, + "learning_rate": 0.00016597512530937378, + "loss": 0.2555, + "step": 32720 + }, + { + "epoch": 0.5105447058089474, + "grad_norm": 4.894479751586914, + "learning_rate": 0.000165964726191219, + "loss": 0.3137, + "step": 32730 + }, + { + "epoch": 0.5107006925812692, + "grad_norm": 0.8396711349487305, + "learning_rate": 0.00016595432707306422, + "loss": 0.3105, + "step": 32740 + }, + { + "epoch": 0.5108566793535908, + "grad_norm": 1.5494674444198608, + "learning_rate": 0.00016594392795490944, + "loss": 0.2271, + "step": 32750 + }, + { + "epoch": 0.5110126661259126, + "grad_norm": 1.3409888744354248, + "learning_rate": 0.00016593352883675466, + "loss": 0.1796, + "step": 32760 + }, + { + "epoch": 0.5111686528982342, + "grad_norm": 1.7590911388397217, + "learning_rate": 0.00016592312971859988, + "loss": 0.2185, + "step": 32770 + }, + { + "epoch": 0.511324639670556, + "grad_norm": 0.7350002527236938, + "learning_rate": 0.0001659127306004451, + "loss": 0.1911, + "step": 32780 + }, + { + "epoch": 0.5114806264428776, + "grad_norm": 0.7176411747932434, + "learning_rate": 0.00016590233148229032, + "loss": 0.6918, + "step": 32790 + }, + { + "epoch": 0.5116366132151994, + "grad_norm": 2.5005321502685547, + "learning_rate": 0.00016589193236413554, + "loss": 0.2329, + "step": 32800 + }, + { + "epoch": 0.511792599987521, + "grad_norm": 1.3916250467300415, + "learning_rate": 0.00016588153324598076, + "loss": 0.2102, + "step": 32810 + }, + { + "epoch": 0.5119485867598428, + "grad_norm": 0.721671998500824, + "learning_rate": 0.00016587113412782598, + "loss": 0.2965, + "step": 32820 + }, + { + "epoch": 0.5121045735321644, + "grad_norm": 0.6120114326477051, + "learning_rate": 0.0001658607350096712, + "loss": 0.2661, + "step": 32830 + }, + { + "epoch": 0.5122605603044862, + "grad_norm": 0.29123347997665405, + "learning_rate": 0.00016585033589151642, + "loss": 0.2455, + "step": 32840 + }, + { + "epoch": 0.5124165470768078, + "grad_norm": 2.7300868034362793, + "learning_rate": 0.00016583993677336164, + "loss": 0.2049, + "step": 32850 + }, + { + "epoch": 0.5125725338491296, + "grad_norm": 1.801928997039795, + "learning_rate": 0.00016582953765520686, + "loss": 0.1725, + "step": 32860 + }, + { + "epoch": 0.5127285206214514, + "grad_norm": 0.6406471729278564, + "learning_rate": 0.00016581913853705208, + "loss": 0.164, + "step": 32870 + }, + { + "epoch": 0.512884507393773, + "grad_norm": 0.9627325534820557, + "learning_rate": 0.0001658087394188973, + "loss": 0.1728, + "step": 32880 + }, + { + "epoch": 0.5130404941660948, + "grad_norm": 1.5446962118148804, + "learning_rate": 0.00016579834030074251, + "loss": 0.2345, + "step": 32890 + }, + { + "epoch": 0.5131964809384164, + "grad_norm": 0.28092706203460693, + "learning_rate": 0.0001657879411825877, + "loss": 0.2124, + "step": 32900 + }, + { + "epoch": 0.5133524677107382, + "grad_norm": 1.7751421928405762, + "learning_rate": 0.00016577754206443295, + "loss": 0.2367, + "step": 32910 + }, + { + "epoch": 0.5135084544830598, + "grad_norm": 1.6964634656906128, + "learning_rate": 0.00016576714294627815, + "loss": 0.2361, + "step": 32920 + }, + { + "epoch": 0.5136644412553816, + "grad_norm": 0.06759823113679886, + "learning_rate": 0.0001657567438281234, + "loss": 0.2329, + "step": 32930 + }, + { + "epoch": 0.5138204280277032, + "grad_norm": 2.2519280910491943, + "learning_rate": 0.00016574634470996859, + "loss": 0.1079, + "step": 32940 + }, + { + "epoch": 0.513976414800025, + "grad_norm": 3.167576551437378, + "learning_rate": 0.00016573594559181383, + "loss": 0.1764, + "step": 32950 + }, + { + "epoch": 0.5141324015723466, + "grad_norm": 1.6734397411346436, + "learning_rate": 0.00016572554647365903, + "loss": 0.2878, + "step": 32960 + }, + { + "epoch": 0.5142883883446684, + "grad_norm": 0.35627374053001404, + "learning_rate": 0.00016571514735550427, + "loss": 0.2337, + "step": 32970 + }, + { + "epoch": 0.51444437511699, + "grad_norm": 1.6321039199829102, + "learning_rate": 0.00016570474823734946, + "loss": 0.2551, + "step": 32980 + }, + { + "epoch": 0.5146003618893118, + "grad_norm": 1.352982997894287, + "learning_rate": 0.0001656943491191947, + "loss": 0.4467, + "step": 32990 + }, + { + "epoch": 0.5147563486616334, + "grad_norm": 0.3558129072189331, + "learning_rate": 0.0001656839500010399, + "loss": 0.1913, + "step": 33000 + }, + { + "epoch": 0.5149123354339552, + "grad_norm": 0.041972286999225616, + "learning_rate": 0.00016567355088288515, + "loss": 0.1704, + "step": 33010 + }, + { + "epoch": 0.515068322206277, + "grad_norm": 1.4168939590454102, + "learning_rate": 0.00016566315176473034, + "loss": 0.4866, + "step": 33020 + }, + { + "epoch": 0.5152243089785986, + "grad_norm": 3.0042929649353027, + "learning_rate": 0.0001656527526465756, + "loss": 0.2473, + "step": 33030 + }, + { + "epoch": 0.5153802957509204, + "grad_norm": 1.027274489402771, + "learning_rate": 0.00016564235352842078, + "loss": 0.436, + "step": 33040 + }, + { + "epoch": 0.515536282523242, + "grad_norm": 1.2330195903778076, + "learning_rate": 0.00016563195441026603, + "loss": 0.4615, + "step": 33050 + }, + { + "epoch": 0.5156922692955638, + "grad_norm": 2.7145066261291504, + "learning_rate": 0.00016562155529211122, + "loss": 0.278, + "step": 33060 + }, + { + "epoch": 0.5158482560678854, + "grad_norm": 1.1273893117904663, + "learning_rate": 0.00016561115617395647, + "loss": 0.3767, + "step": 33070 + }, + { + "epoch": 0.5160042428402072, + "grad_norm": 2.466265916824341, + "learning_rate": 0.00016560075705580166, + "loss": 0.1879, + "step": 33080 + }, + { + "epoch": 0.5161602296125288, + "grad_norm": 0.3759515583515167, + "learning_rate": 0.0001655903579376469, + "loss": 0.4004, + "step": 33090 + }, + { + "epoch": 0.5163162163848506, + "grad_norm": 1.8821336030960083, + "learning_rate": 0.0001655799588194921, + "loss": 0.1622, + "step": 33100 + }, + { + "epoch": 0.5164722031571722, + "grad_norm": 0.8162096738815308, + "learning_rate": 0.00016556955970133735, + "loss": 0.1341, + "step": 33110 + }, + { + "epoch": 0.516628189929494, + "grad_norm": 0.008498801849782467, + "learning_rate": 0.00016555916058318254, + "loss": 0.2546, + "step": 33120 + }, + { + "epoch": 0.5167841767018156, + "grad_norm": 0.17536675930023193, + "learning_rate": 0.00016554876146502779, + "loss": 0.2934, + "step": 33130 + }, + { + "epoch": 0.5169401634741374, + "grad_norm": 3.539898633956909, + "learning_rate": 0.00016553836234687298, + "loss": 0.4959, + "step": 33140 + }, + { + "epoch": 0.517096150246459, + "grad_norm": 2.190887212753296, + "learning_rate": 0.00016552796322871823, + "loss": 0.2914, + "step": 33150 + }, + { + "epoch": 0.5172521370187808, + "grad_norm": 1.6388943195343018, + "learning_rate": 0.00016551756411056342, + "loss": 0.1317, + "step": 33160 + }, + { + "epoch": 0.5174081237911026, + "grad_norm": 0.33063387870788574, + "learning_rate": 0.00016550716499240866, + "loss": 0.4258, + "step": 33170 + }, + { + "epoch": 0.5175641105634242, + "grad_norm": 0.5168288946151733, + "learning_rate": 0.00016549676587425386, + "loss": 0.2548, + "step": 33180 + }, + { + "epoch": 0.517720097335746, + "grad_norm": 3.9570281505584717, + "learning_rate": 0.0001654863667560991, + "loss": 0.3179, + "step": 33190 + }, + { + "epoch": 0.5178760841080676, + "grad_norm": 1.127255916595459, + "learning_rate": 0.0001654759676379443, + "loss": 0.233, + "step": 33200 + }, + { + "epoch": 0.5180320708803894, + "grad_norm": 0.36299410462379456, + "learning_rate": 0.00016546556851978954, + "loss": 0.4614, + "step": 33210 + }, + { + "epoch": 0.518188057652711, + "grad_norm": 2.914668321609497, + "learning_rate": 0.00016545516940163474, + "loss": 0.1603, + "step": 33220 + }, + { + "epoch": 0.5183440444250328, + "grad_norm": 6.991094589233398, + "learning_rate": 0.00016544477028347998, + "loss": 0.2259, + "step": 33230 + }, + { + "epoch": 0.5185000311973544, + "grad_norm": 2.0594241619110107, + "learning_rate": 0.00016543437116532517, + "loss": 0.1885, + "step": 33240 + }, + { + "epoch": 0.5186560179696762, + "grad_norm": 0.5594693422317505, + "learning_rate": 0.00016542397204717042, + "loss": 0.4333, + "step": 33250 + }, + { + "epoch": 0.5188120047419978, + "grad_norm": 4.314854621887207, + "learning_rate": 0.00016541357292901561, + "loss": 0.2103, + "step": 33260 + }, + { + "epoch": 0.5189679915143196, + "grad_norm": 0.5609000325202942, + "learning_rate": 0.00016540317381086086, + "loss": 0.2498, + "step": 33270 + }, + { + "epoch": 0.5191239782866413, + "grad_norm": 3.680088758468628, + "learning_rate": 0.00016539277469270605, + "loss": 0.3465, + "step": 33280 + }, + { + "epoch": 0.519279965058963, + "grad_norm": 1.4993672370910645, + "learning_rate": 0.0001653823755745513, + "loss": 0.3272, + "step": 33290 + }, + { + "epoch": 0.5194359518312847, + "grad_norm": 0.18005084991455078, + "learning_rate": 0.0001653719764563965, + "loss": 0.1706, + "step": 33300 + }, + { + "epoch": 0.5195919386036064, + "grad_norm": 0.5105524659156799, + "learning_rate": 0.00016536157733824174, + "loss": 0.0937, + "step": 33310 + }, + { + "epoch": 0.5197479253759281, + "grad_norm": 1.2806589603424072, + "learning_rate": 0.00016535117822008693, + "loss": 0.1578, + "step": 33320 + }, + { + "epoch": 0.5199039121482498, + "grad_norm": 0.22015990316867828, + "learning_rate": 0.00016534077910193218, + "loss": 0.2081, + "step": 33330 + }, + { + "epoch": 0.5200598989205716, + "grad_norm": 0.34638991951942444, + "learning_rate": 0.00016533037998377737, + "loss": 0.2389, + "step": 33340 + }, + { + "epoch": 0.5202158856928932, + "grad_norm": 0.8709302544593811, + "learning_rate": 0.0001653199808656226, + "loss": 0.2112, + "step": 33350 + }, + { + "epoch": 0.520371872465215, + "grad_norm": 1.2309521436691284, + "learning_rate": 0.0001653095817474678, + "loss": 0.3448, + "step": 33360 + }, + { + "epoch": 0.5205278592375366, + "grad_norm": 1.3155115842819214, + "learning_rate": 0.00016529918262931303, + "loss": 0.1572, + "step": 33370 + }, + { + "epoch": 0.5206838460098584, + "grad_norm": 0.358751118183136, + "learning_rate": 0.00016528878351115825, + "loss": 0.3011, + "step": 33380 + }, + { + "epoch": 0.52083983278218, + "grad_norm": 3.800628662109375, + "learning_rate": 0.00016527838439300347, + "loss": 0.458, + "step": 33390 + }, + { + "epoch": 0.5209958195545018, + "grad_norm": 1.1253124475479126, + "learning_rate": 0.0001652679852748487, + "loss": 0.1238, + "step": 33400 + }, + { + "epoch": 0.5211518063268235, + "grad_norm": 0.46979108452796936, + "learning_rate": 0.0001652575861566939, + "loss": 0.0745, + "step": 33410 + }, + { + "epoch": 0.5213077930991452, + "grad_norm": 7.921492576599121, + "learning_rate": 0.00016524718703853913, + "loss": 0.4004, + "step": 33420 + }, + { + "epoch": 0.5214637798714669, + "grad_norm": 1.643127679824829, + "learning_rate": 0.00016523678792038435, + "loss": 0.2858, + "step": 33430 + }, + { + "epoch": 0.5216197666437886, + "grad_norm": 7.096631050109863, + "learning_rate": 0.0001652263888022296, + "loss": 0.4735, + "step": 33440 + }, + { + "epoch": 0.5217757534161103, + "grad_norm": 0.7280219197273254, + "learning_rate": 0.0001652159896840748, + "loss": 0.1841, + "step": 33450 + }, + { + "epoch": 0.521931740188432, + "grad_norm": 0.9415522813796997, + "learning_rate": 0.00016520559056592003, + "loss": 0.2169, + "step": 33460 + }, + { + "epoch": 0.5220877269607537, + "grad_norm": 2.264862060546875, + "learning_rate": 0.00016519519144776523, + "loss": 0.1902, + "step": 33470 + }, + { + "epoch": 0.5222437137330754, + "grad_norm": 2.4566822052001953, + "learning_rate": 0.00016518479232961047, + "loss": 0.3329, + "step": 33480 + }, + { + "epoch": 0.5223997005053972, + "grad_norm": 0.7330378890037537, + "learning_rate": 0.00016517439321145567, + "loss": 0.2341, + "step": 33490 + }, + { + "epoch": 0.5225556872777188, + "grad_norm": 1.6409542560577393, + "learning_rate": 0.0001651639940933009, + "loss": 0.3077, + "step": 33500 + }, + { + "epoch": 0.5227116740500406, + "grad_norm": 1.0647432804107666, + "learning_rate": 0.0001651535949751461, + "loss": 0.1935, + "step": 33510 + }, + { + "epoch": 0.5228676608223622, + "grad_norm": 0.507420539855957, + "learning_rate": 0.00016514319585699135, + "loss": 0.1314, + "step": 33520 + }, + { + "epoch": 0.523023647594684, + "grad_norm": 1.0064164400100708, + "learning_rate": 0.00016513279673883654, + "loss": 0.344, + "step": 33530 + }, + { + "epoch": 0.5231796343670057, + "grad_norm": 1.1936209201812744, + "learning_rate": 0.0001651223976206818, + "loss": 0.1948, + "step": 33540 + }, + { + "epoch": 0.5233356211393274, + "grad_norm": 0.04992926865816116, + "learning_rate": 0.00016511199850252698, + "loss": 0.1363, + "step": 33550 + }, + { + "epoch": 0.5234916079116491, + "grad_norm": 1.8256275653839111, + "learning_rate": 0.00016510159938437223, + "loss": 0.443, + "step": 33560 + }, + { + "epoch": 0.5236475946839708, + "grad_norm": 4.914908409118652, + "learning_rate": 0.00016509120026621742, + "loss": 0.4654, + "step": 33570 + }, + { + "epoch": 0.5238035814562925, + "grad_norm": 1.1625486612319946, + "learning_rate": 0.00016508080114806267, + "loss": 0.1974, + "step": 33580 + }, + { + "epoch": 0.5239595682286142, + "grad_norm": 6.6192522048950195, + "learning_rate": 0.00016507040202990786, + "loss": 0.2131, + "step": 33590 + }, + { + "epoch": 0.5241155550009359, + "grad_norm": 0.7369065880775452, + "learning_rate": 0.0001650600029117531, + "loss": 0.2371, + "step": 33600 + }, + { + "epoch": 0.5242715417732576, + "grad_norm": 1.5238152742385864, + "learning_rate": 0.0001650496037935983, + "loss": 0.1557, + "step": 33610 + }, + { + "epoch": 0.5244275285455793, + "grad_norm": 1.0418007373809814, + "learning_rate": 0.00016503920467544355, + "loss": 0.1878, + "step": 33620 + }, + { + "epoch": 0.524583515317901, + "grad_norm": 0.790117084980011, + "learning_rate": 0.00016502880555728874, + "loss": 0.2195, + "step": 33630 + }, + { + "epoch": 0.5247395020902228, + "grad_norm": 1.6712257862091064, + "learning_rate": 0.000165018406439134, + "loss": 0.1602, + "step": 33640 + }, + { + "epoch": 0.5248954888625444, + "grad_norm": 0.19236230850219727, + "learning_rate": 0.00016500800732097918, + "loss": 0.2526, + "step": 33650 + }, + { + "epoch": 0.5250514756348662, + "grad_norm": 1.3519701957702637, + "learning_rate": 0.00016499760820282443, + "loss": 0.4686, + "step": 33660 + }, + { + "epoch": 0.5252074624071879, + "grad_norm": 1.694342851638794, + "learning_rate": 0.00016498720908466962, + "loss": 0.1859, + "step": 33670 + }, + { + "epoch": 0.5253634491795096, + "grad_norm": 5.225239276885986, + "learning_rate": 0.00016497680996651487, + "loss": 0.2944, + "step": 33680 + }, + { + "epoch": 0.5255194359518313, + "grad_norm": 2.0208842754364014, + "learning_rate": 0.00016496641084836006, + "loss": 0.2421, + "step": 33690 + }, + { + "epoch": 0.525675422724153, + "grad_norm": 0.7954996228218079, + "learning_rate": 0.0001649560117302053, + "loss": 0.2853, + "step": 33700 + }, + { + "epoch": 0.5258314094964747, + "grad_norm": 2.296086072921753, + "learning_rate": 0.0001649456126120505, + "loss": 0.1333, + "step": 33710 + }, + { + "epoch": 0.5259873962687964, + "grad_norm": 1.1779128313064575, + "learning_rate": 0.00016493521349389574, + "loss": 0.1066, + "step": 33720 + }, + { + "epoch": 0.5261433830411181, + "grad_norm": 0.1756065934896469, + "learning_rate": 0.00016492481437574094, + "loss": 0.1352, + "step": 33730 + }, + { + "epoch": 0.5262993698134398, + "grad_norm": 0.13100725412368774, + "learning_rate": 0.00016491441525758618, + "loss": 0.2399, + "step": 33740 + }, + { + "epoch": 0.5264553565857615, + "grad_norm": 5.532008171081543, + "learning_rate": 0.00016490401613943138, + "loss": 0.2896, + "step": 33750 + }, + { + "epoch": 0.5266113433580832, + "grad_norm": 1.319886565208435, + "learning_rate": 0.00016489361702127662, + "loss": 0.3275, + "step": 33760 + }, + { + "epoch": 0.5267673301304049, + "grad_norm": 1.5550974607467651, + "learning_rate": 0.00016488321790312182, + "loss": 0.2677, + "step": 33770 + }, + { + "epoch": 0.5269233169027266, + "grad_norm": 1.8936737775802612, + "learning_rate": 0.00016487281878496706, + "loss": 0.1955, + "step": 33780 + }, + { + "epoch": 0.5270793036750484, + "grad_norm": 0.3653401732444763, + "learning_rate": 0.00016486241966681226, + "loss": 0.0723, + "step": 33790 + }, + { + "epoch": 0.52723529044737, + "grad_norm": 2.861341714859009, + "learning_rate": 0.00016485202054865747, + "loss": 0.2412, + "step": 33800 + }, + { + "epoch": 0.5273912772196918, + "grad_norm": 1.5291428565979004, + "learning_rate": 0.0001648416214305027, + "loss": 0.0871, + "step": 33810 + }, + { + "epoch": 0.5275472639920135, + "grad_norm": 1.0372581481933594, + "learning_rate": 0.00016483122231234791, + "loss": 0.4705, + "step": 33820 + }, + { + "epoch": 0.5277032507643352, + "grad_norm": 1.1943141222000122, + "learning_rate": 0.00016482082319419313, + "loss": 0.2848, + "step": 33830 + }, + { + "epoch": 0.5278592375366569, + "grad_norm": 1.9008225202560425, + "learning_rate": 0.00016481042407603835, + "loss": 0.139, + "step": 33840 + }, + { + "epoch": 0.5280152243089786, + "grad_norm": 2.132089138031006, + "learning_rate": 0.00016480002495788357, + "loss": 0.2119, + "step": 33850 + }, + { + "epoch": 0.5281712110813003, + "grad_norm": 0.24524426460266113, + "learning_rate": 0.0001647896258397288, + "loss": 0.1643, + "step": 33860 + }, + { + "epoch": 0.528327197853622, + "grad_norm": 1.6469637155532837, + "learning_rate": 0.000164779226721574, + "loss": 0.363, + "step": 33870 + }, + { + "epoch": 0.5284831846259437, + "grad_norm": 0.8767328858375549, + "learning_rate": 0.00016476882760341923, + "loss": 0.2632, + "step": 33880 + }, + { + "epoch": 0.5286391713982654, + "grad_norm": 0.06347586214542389, + "learning_rate": 0.00016475842848526445, + "loss": 0.3204, + "step": 33890 + }, + { + "epoch": 0.5287951581705871, + "grad_norm": 0.09782540798187256, + "learning_rate": 0.00016474802936710967, + "loss": 0.2211, + "step": 33900 + }, + { + "epoch": 0.5289511449429088, + "grad_norm": 3.2998859882354736, + "learning_rate": 0.0001647376302489549, + "loss": 0.265, + "step": 33910 + }, + { + "epoch": 0.5291071317152305, + "grad_norm": 0.43594226241111755, + "learning_rate": 0.0001647272311308001, + "loss": 0.194, + "step": 33920 + }, + { + "epoch": 0.5292631184875523, + "grad_norm": 1.5166605710983276, + "learning_rate": 0.00016471683201264533, + "loss": 0.2675, + "step": 33930 + }, + { + "epoch": 0.529419105259874, + "grad_norm": 0.6056640148162842, + "learning_rate": 0.00016470643289449055, + "loss": 0.0576, + "step": 33940 + }, + { + "epoch": 0.5295750920321957, + "grad_norm": 0.25410348176956177, + "learning_rate": 0.00016469603377633577, + "loss": 0.1124, + "step": 33950 + }, + { + "epoch": 0.5297310788045174, + "grad_norm": 1.770642876625061, + "learning_rate": 0.000164685634658181, + "loss": 0.3295, + "step": 33960 + }, + { + "epoch": 0.5298870655768391, + "grad_norm": 0.0607205331325531, + "learning_rate": 0.0001646752355400262, + "loss": 0.2369, + "step": 33970 + }, + { + "epoch": 0.5300430523491608, + "grad_norm": 0.5557095408439636, + "learning_rate": 0.00016466483642187143, + "loss": 0.1681, + "step": 33980 + }, + { + "epoch": 0.5301990391214825, + "grad_norm": 0.5192957520484924, + "learning_rate": 0.00016465443730371665, + "loss": 0.2649, + "step": 33990 + }, + { + "epoch": 0.5303550258938042, + "grad_norm": 0.04804835096001625, + "learning_rate": 0.00016464403818556187, + "loss": 0.2639, + "step": 34000 + }, + { + "epoch": 0.5305110126661259, + "grad_norm": 0.02673129364848137, + "learning_rate": 0.0001646336390674071, + "loss": 0.169, + "step": 34010 + }, + { + "epoch": 0.5306669994384476, + "grad_norm": 1.0084244012832642, + "learning_rate": 0.0001646232399492523, + "loss": 0.353, + "step": 34020 + }, + { + "epoch": 0.5308229862107693, + "grad_norm": 2.2202091217041016, + "learning_rate": 0.00016461284083109753, + "loss": 0.3206, + "step": 34030 + }, + { + "epoch": 0.530978972983091, + "grad_norm": 0.5573744773864746, + "learning_rate": 0.00016460244171294275, + "loss": 0.231, + "step": 34040 + }, + { + "epoch": 0.5311349597554127, + "grad_norm": 0.39700084924697876, + "learning_rate": 0.00016459204259478797, + "loss": 0.1412, + "step": 34050 + }, + { + "epoch": 0.5312909465277345, + "grad_norm": 2.582963228225708, + "learning_rate": 0.00016458164347663319, + "loss": 0.3334, + "step": 34060 + }, + { + "epoch": 0.5314469333000561, + "grad_norm": 0.22781169414520264, + "learning_rate": 0.0001645712443584784, + "loss": 0.2767, + "step": 34070 + }, + { + "epoch": 0.5316029200723779, + "grad_norm": 1.2051042318344116, + "learning_rate": 0.00016456084524032362, + "loss": 0.2045, + "step": 34080 + }, + { + "epoch": 0.5317589068446996, + "grad_norm": 0.42760100960731506, + "learning_rate": 0.00016455044612216884, + "loss": 0.1151, + "step": 34090 + }, + { + "epoch": 0.5319148936170213, + "grad_norm": 0.14440476894378662, + "learning_rate": 0.00016454004700401406, + "loss": 0.2386, + "step": 34100 + }, + { + "epoch": 0.532070880389343, + "grad_norm": 2.2777981758117676, + "learning_rate": 0.00016452964788585928, + "loss": 0.2009, + "step": 34110 + }, + { + "epoch": 0.5322268671616647, + "grad_norm": 0.9206979274749756, + "learning_rate": 0.0001645192487677045, + "loss": 0.2745, + "step": 34120 + }, + { + "epoch": 0.5323828539339864, + "grad_norm": 1.6947574615478516, + "learning_rate": 0.00016450884964954972, + "loss": 0.2584, + "step": 34130 + }, + { + "epoch": 0.5325388407063081, + "grad_norm": 0.401444673538208, + "learning_rate": 0.00016449845053139494, + "loss": 0.2218, + "step": 34140 + }, + { + "epoch": 0.5326948274786298, + "grad_norm": 0.08261553198099136, + "learning_rate": 0.00016448805141324016, + "loss": 0.2775, + "step": 34150 + }, + { + "epoch": 0.5328508142509515, + "grad_norm": 0.1017974391579628, + "learning_rate": 0.00016447765229508538, + "loss": 0.2095, + "step": 34160 + }, + { + "epoch": 0.5330068010232732, + "grad_norm": 1.3759571313858032, + "learning_rate": 0.0001644672531769306, + "loss": 0.2643, + "step": 34170 + }, + { + "epoch": 0.5331627877955949, + "grad_norm": 1.2654389142990112, + "learning_rate": 0.00016445685405877582, + "loss": 0.2949, + "step": 34180 + }, + { + "epoch": 0.5333187745679167, + "grad_norm": 1.5481843948364258, + "learning_rate": 0.00016444645494062104, + "loss": 0.1264, + "step": 34190 + }, + { + "epoch": 0.5334747613402383, + "grad_norm": 1.8094528913497925, + "learning_rate": 0.00016443605582246626, + "loss": 0.2727, + "step": 34200 + }, + { + "epoch": 0.5336307481125601, + "grad_norm": 2.224538564682007, + "learning_rate": 0.00016442565670431148, + "loss": 0.3096, + "step": 34210 + }, + { + "epoch": 0.5337867348848817, + "grad_norm": 0.6375226974487305, + "learning_rate": 0.0001644152575861567, + "loss": 0.2251, + "step": 34220 + }, + { + "epoch": 0.5339427216572035, + "grad_norm": 3.727106809616089, + "learning_rate": 0.00016440485846800192, + "loss": 0.4374, + "step": 34230 + }, + { + "epoch": 0.5340987084295252, + "grad_norm": 0.13345426321029663, + "learning_rate": 0.00016439445934984714, + "loss": 0.2011, + "step": 34240 + }, + { + "epoch": 0.5342546952018469, + "grad_norm": 2.1658668518066406, + "learning_rate": 0.00016438406023169236, + "loss": 0.2457, + "step": 34250 + }, + { + "epoch": 0.5344106819741686, + "grad_norm": 19.238407135009766, + "learning_rate": 0.00016437366111353758, + "loss": 0.2756, + "step": 34260 + }, + { + "epoch": 0.5345666687464903, + "grad_norm": 1.0292778015136719, + "learning_rate": 0.0001643632619953828, + "loss": 0.1646, + "step": 34270 + }, + { + "epoch": 0.534722655518812, + "grad_norm": 0.9372987747192383, + "learning_rate": 0.00016435286287722802, + "loss": 0.2762, + "step": 34280 + }, + { + "epoch": 0.5348786422911337, + "grad_norm": 0.3918002247810364, + "learning_rate": 0.00016434246375907324, + "loss": 0.1406, + "step": 34290 + }, + { + "epoch": 0.5350346290634554, + "grad_norm": 1.3518732786178589, + "learning_rate": 0.00016433206464091846, + "loss": 0.2992, + "step": 34300 + }, + { + "epoch": 0.5351906158357771, + "grad_norm": 0.73117595911026, + "learning_rate": 0.00016432166552276368, + "loss": 0.1172, + "step": 34310 + }, + { + "epoch": 0.5353466026080989, + "grad_norm": 0.06655958294868469, + "learning_rate": 0.0001643112664046089, + "loss": 0.1332, + "step": 34320 + }, + { + "epoch": 0.5355025893804205, + "grad_norm": 3.2587168216705322, + "learning_rate": 0.00016430086728645412, + "loss": 0.2052, + "step": 34330 + }, + { + "epoch": 0.5356585761527423, + "grad_norm": 2.6668431758880615, + "learning_rate": 0.00016429046816829934, + "loss": 0.1295, + "step": 34340 + }, + { + "epoch": 0.5358145629250639, + "grad_norm": 1.316080927848816, + "learning_rate": 0.00016428006905014456, + "loss": 0.2072, + "step": 34350 + }, + { + "epoch": 0.5359705496973857, + "grad_norm": 5.144528388977051, + "learning_rate": 0.00016426966993198977, + "loss": 0.1934, + "step": 34360 + }, + { + "epoch": 0.5361265364697073, + "grad_norm": 1.3529599905014038, + "learning_rate": 0.000164259270813835, + "loss": 0.2077, + "step": 34370 + }, + { + "epoch": 0.5362825232420291, + "grad_norm": 0.9132925868034363, + "learning_rate": 0.00016424887169568021, + "loss": 0.2404, + "step": 34380 + }, + { + "epoch": 0.5364385100143508, + "grad_norm": 1.520033597946167, + "learning_rate": 0.00016423847257752543, + "loss": 0.1765, + "step": 34390 + }, + { + "epoch": 0.5365944967866725, + "grad_norm": 2.343975305557251, + "learning_rate": 0.00016422807345937065, + "loss": 0.1959, + "step": 34400 + }, + { + "epoch": 0.5367504835589942, + "grad_norm": 0.5855118632316589, + "learning_rate": 0.00016421767434121587, + "loss": 0.0848, + "step": 34410 + }, + { + "epoch": 0.5369064703313159, + "grad_norm": 1.9108648300170898, + "learning_rate": 0.0001642072752230611, + "loss": 0.2406, + "step": 34420 + }, + { + "epoch": 0.5370624571036376, + "grad_norm": 1.0323792695999146, + "learning_rate": 0.0001641968761049063, + "loss": 0.2298, + "step": 34430 + }, + { + "epoch": 0.5372184438759593, + "grad_norm": 0.17435620725154877, + "learning_rate": 0.00016418647698675153, + "loss": 0.1396, + "step": 34440 + }, + { + "epoch": 0.537374430648281, + "grad_norm": 0.3526577353477478, + "learning_rate": 0.00016417607786859675, + "loss": 0.172, + "step": 34450 + }, + { + "epoch": 0.5375304174206027, + "grad_norm": 2.73685884475708, + "learning_rate": 0.00016416567875044197, + "loss": 0.1167, + "step": 34460 + }, + { + "epoch": 0.5376864041929245, + "grad_norm": 2.765693426132202, + "learning_rate": 0.0001641552796322872, + "loss": 0.1356, + "step": 34470 + }, + { + "epoch": 0.5378423909652461, + "grad_norm": 0.47628021240234375, + "learning_rate": 0.0001641448805141324, + "loss": 0.3104, + "step": 34480 + }, + { + "epoch": 0.5379983777375679, + "grad_norm": 1.1307590007781982, + "learning_rate": 0.00016413448139597763, + "loss": 0.3247, + "step": 34490 + }, + { + "epoch": 0.5381543645098895, + "grad_norm": 0.07136381417512894, + "learning_rate": 0.00016412408227782285, + "loss": 0.319, + "step": 34500 + }, + { + "epoch": 0.5383103512822113, + "grad_norm": 1.7850221395492554, + "learning_rate": 0.00016411368315966807, + "loss": 0.2515, + "step": 34510 + }, + { + "epoch": 0.5384663380545329, + "grad_norm": 1.3870742321014404, + "learning_rate": 0.0001641032840415133, + "loss": 0.256, + "step": 34520 + }, + { + "epoch": 0.5386223248268547, + "grad_norm": 1.1329221725463867, + "learning_rate": 0.0001640928849233585, + "loss": 0.2255, + "step": 34530 + }, + { + "epoch": 0.5387783115991764, + "grad_norm": 0.23236137628555298, + "learning_rate": 0.00016408248580520373, + "loss": 0.4053, + "step": 34540 + }, + { + "epoch": 0.5389342983714981, + "grad_norm": 3.350924253463745, + "learning_rate": 0.00016407208668704895, + "loss": 0.2816, + "step": 34550 + }, + { + "epoch": 0.5390902851438198, + "grad_norm": 1.7459170818328857, + "learning_rate": 0.00016406168756889417, + "loss": 0.0834, + "step": 34560 + }, + { + "epoch": 0.5392462719161415, + "grad_norm": 0.89351487159729, + "learning_rate": 0.0001640512884507394, + "loss": 0.1146, + "step": 34570 + }, + { + "epoch": 0.5394022586884633, + "grad_norm": 0.49410831928253174, + "learning_rate": 0.0001640408893325846, + "loss": 0.2319, + "step": 34580 + }, + { + "epoch": 0.5395582454607849, + "grad_norm": 0.009971237741410732, + "learning_rate": 0.00016403049021442983, + "loss": 0.1674, + "step": 34590 + }, + { + "epoch": 0.5397142322331067, + "grad_norm": 1.420511245727539, + "learning_rate": 0.00016402009109627505, + "loss": 0.1469, + "step": 34600 + }, + { + "epoch": 0.5398702190054283, + "grad_norm": 2.5686769485473633, + "learning_rate": 0.00016400969197812027, + "loss": 0.2607, + "step": 34610 + }, + { + "epoch": 0.5400262057777501, + "grad_norm": 2.5808913707733154, + "learning_rate": 0.00016399929285996549, + "loss": 0.4412, + "step": 34620 + }, + { + "epoch": 0.5401821925500717, + "grad_norm": 0.2778591811656952, + "learning_rate": 0.0001639888937418107, + "loss": 0.1563, + "step": 34630 + }, + { + "epoch": 0.5403381793223935, + "grad_norm": 1.6448099613189697, + "learning_rate": 0.00016397849462365592, + "loss": 0.2484, + "step": 34640 + }, + { + "epoch": 0.5404941660947151, + "grad_norm": 0.003213417250663042, + "learning_rate": 0.00016396809550550114, + "loss": 0.1582, + "step": 34650 + }, + { + "epoch": 0.5406501528670369, + "grad_norm": 2.3433990478515625, + "learning_rate": 0.00016395769638734636, + "loss": 0.2774, + "step": 34660 + }, + { + "epoch": 0.5408061396393585, + "grad_norm": 0.09544426947832108, + "learning_rate": 0.00016394729726919158, + "loss": 0.1242, + "step": 34670 + }, + { + "epoch": 0.5409621264116803, + "grad_norm": 1.8934235572814941, + "learning_rate": 0.0001639368981510368, + "loss": 0.1636, + "step": 34680 + }, + { + "epoch": 0.541118113184002, + "grad_norm": 2.8528554439544678, + "learning_rate": 0.00016392649903288202, + "loss": 0.2524, + "step": 34690 + }, + { + "epoch": 0.5412740999563237, + "grad_norm": 2.837836980819702, + "learning_rate": 0.00016391609991472722, + "loss": 0.1808, + "step": 34700 + }, + { + "epoch": 0.5414300867286455, + "grad_norm": 4.349740028381348, + "learning_rate": 0.00016390570079657246, + "loss": 0.12, + "step": 34710 + }, + { + "epoch": 0.5415860735009671, + "grad_norm": 0.6011945605278015, + "learning_rate": 0.00016389530167841765, + "loss": 0.2865, + "step": 34720 + }, + { + "epoch": 0.5417420602732889, + "grad_norm": 1.2079488039016724, + "learning_rate": 0.0001638849025602629, + "loss": 0.125, + "step": 34730 + }, + { + "epoch": 0.5418980470456105, + "grad_norm": 1.9549164772033691, + "learning_rate": 0.0001638745034421081, + "loss": 0.3285, + "step": 34740 + }, + { + "epoch": 0.5420540338179323, + "grad_norm": 0.6969407200813293, + "learning_rate": 0.00016386410432395334, + "loss": 0.3267, + "step": 34750 + }, + { + "epoch": 0.5422100205902539, + "grad_norm": 1.0561965703964233, + "learning_rate": 0.00016385370520579853, + "loss": 0.1463, + "step": 34760 + }, + { + "epoch": 0.5423660073625757, + "grad_norm": 1.8208931684494019, + "learning_rate": 0.00016384330608764378, + "loss": 0.1713, + "step": 34770 + }, + { + "epoch": 0.5425219941348973, + "grad_norm": 0.6349910497665405, + "learning_rate": 0.000163832906969489, + "loss": 0.2244, + "step": 34780 + }, + { + "epoch": 0.5426779809072191, + "grad_norm": 1.1976515054702759, + "learning_rate": 0.00016382250785133422, + "loss": 0.3407, + "step": 34790 + }, + { + "epoch": 0.5428339676795407, + "grad_norm": 0.5970319509506226, + "learning_rate": 0.00016381210873317944, + "loss": 0.1419, + "step": 34800 + }, + { + "epoch": 0.5429899544518625, + "grad_norm": 0.14623159170150757, + "learning_rate": 0.00016380170961502466, + "loss": 0.14, + "step": 34810 + }, + { + "epoch": 0.5431459412241841, + "grad_norm": 1.4332351684570312, + "learning_rate": 0.00016379131049686988, + "loss": 0.1093, + "step": 34820 + }, + { + "epoch": 0.5433019279965059, + "grad_norm": 1.1443063020706177, + "learning_rate": 0.0001637809113787151, + "loss": 0.3676, + "step": 34830 + }, + { + "epoch": 0.5434579147688277, + "grad_norm": 1.3897461891174316, + "learning_rate": 0.00016377051226056032, + "loss": 0.3713, + "step": 34840 + }, + { + "epoch": 0.5436139015411493, + "grad_norm": 1.7773199081420898, + "learning_rate": 0.00016376011314240554, + "loss": 0.2249, + "step": 34850 + }, + { + "epoch": 0.5437698883134711, + "grad_norm": 0.8524389266967773, + "learning_rate": 0.00016374971402425076, + "loss": 0.2158, + "step": 34860 + }, + { + "epoch": 0.5439258750857927, + "grad_norm": 0.5572389960289001, + "learning_rate": 0.00016373931490609598, + "loss": 0.215, + "step": 34870 + }, + { + "epoch": 0.5440818618581145, + "grad_norm": 3.0238709449768066, + "learning_rate": 0.0001637289157879412, + "loss": 0.3765, + "step": 34880 + }, + { + "epoch": 0.5442378486304361, + "grad_norm": 1.1655421257019043, + "learning_rate": 0.00016371851666978642, + "loss": 0.2491, + "step": 34890 + }, + { + "epoch": 0.5443938354027579, + "grad_norm": 0.49492090940475464, + "learning_rate": 0.00016370811755163164, + "loss": 0.1287, + "step": 34900 + }, + { + "epoch": 0.5445498221750795, + "grad_norm": 0.2732921540737152, + "learning_rate": 0.00016369771843347685, + "loss": 0.2154, + "step": 34910 + }, + { + "epoch": 0.5447058089474013, + "grad_norm": 2.5807769298553467, + "learning_rate": 0.00016368731931532207, + "loss": 0.2032, + "step": 34920 + }, + { + "epoch": 0.5448617957197229, + "grad_norm": 0.5044315457344055, + "learning_rate": 0.0001636769201971673, + "loss": 0.2595, + "step": 34930 + }, + { + "epoch": 0.5450177824920447, + "grad_norm": 2.1248972415924072, + "learning_rate": 0.00016366652107901251, + "loss": 0.5651, + "step": 34940 + }, + { + "epoch": 0.5451737692643663, + "grad_norm": 0.8391468524932861, + "learning_rate": 0.00016365612196085773, + "loss": 0.3028, + "step": 34950 + }, + { + "epoch": 0.5453297560366881, + "grad_norm": 0.36081477999687195, + "learning_rate": 0.00016364572284270295, + "loss": 0.1073, + "step": 34960 + }, + { + "epoch": 0.5454857428090097, + "grad_norm": 3.175804376602173, + "learning_rate": 0.00016363532372454817, + "loss": 0.1889, + "step": 34970 + }, + { + "epoch": 0.5456417295813315, + "grad_norm": 0.20647937059402466, + "learning_rate": 0.0001636249246063934, + "loss": 0.2416, + "step": 34980 + }, + { + "epoch": 0.5457977163536533, + "grad_norm": 2.3525943756103516, + "learning_rate": 0.0001636145254882386, + "loss": 0.3179, + "step": 34990 + }, + { + "epoch": 0.5459537031259749, + "grad_norm": 0.2641216814517975, + "learning_rate": 0.00016360412637008383, + "loss": 0.2157, + "step": 35000 + }, + { + "epoch": 0.5461096898982967, + "grad_norm": 0.09906073659658432, + "learning_rate": 0.00016359372725192905, + "loss": 0.2576, + "step": 35010 + }, + { + "epoch": 0.5462656766706183, + "grad_norm": 0.13971920311450958, + "learning_rate": 0.00016358332813377427, + "loss": 0.1298, + "step": 35020 + }, + { + "epoch": 0.5464216634429401, + "grad_norm": 1.4535638093948364, + "learning_rate": 0.0001635729290156195, + "loss": 0.2357, + "step": 35030 + }, + { + "epoch": 0.5465776502152617, + "grad_norm": 0.9107828736305237, + "learning_rate": 0.0001635625298974647, + "loss": 0.3049, + "step": 35040 + }, + { + "epoch": 0.5467336369875835, + "grad_norm": 0.0377386212348938, + "learning_rate": 0.00016355213077930993, + "loss": 0.2066, + "step": 35050 + }, + { + "epoch": 0.5468896237599051, + "grad_norm": 2.7800869941711426, + "learning_rate": 0.00016354173166115515, + "loss": 0.2974, + "step": 35060 + }, + { + "epoch": 0.5470456105322269, + "grad_norm": 1.6247998476028442, + "learning_rate": 0.00016353133254300037, + "loss": 0.3822, + "step": 35070 + }, + { + "epoch": 0.5472015973045485, + "grad_norm": 1.6479015350341797, + "learning_rate": 0.0001635209334248456, + "loss": 0.3401, + "step": 35080 + }, + { + "epoch": 0.5473575840768703, + "grad_norm": 0.1879737675189972, + "learning_rate": 0.0001635105343066908, + "loss": 0.1549, + "step": 35090 + }, + { + "epoch": 0.5475135708491919, + "grad_norm": 0.39355361461639404, + "learning_rate": 0.00016350013518853603, + "loss": 0.2032, + "step": 35100 + }, + { + "epoch": 0.5476695576215137, + "grad_norm": 2.9976983070373535, + "learning_rate": 0.00016348973607038125, + "loss": 0.3867, + "step": 35110 + }, + { + "epoch": 0.5478255443938354, + "grad_norm": 0.8242707252502441, + "learning_rate": 0.00016347933695222647, + "loss": 0.3601, + "step": 35120 + }, + { + "epoch": 0.5479815311661571, + "grad_norm": 0.874955415725708, + "learning_rate": 0.0001634689378340717, + "loss": 0.2709, + "step": 35130 + }, + { + "epoch": 0.5481375179384789, + "grad_norm": 0.05980971083045006, + "learning_rate": 0.0001634585387159169, + "loss": 0.2646, + "step": 35140 + }, + { + "epoch": 0.5482935047108005, + "grad_norm": 1.8437328338623047, + "learning_rate": 0.0001634481395977621, + "loss": 0.1965, + "step": 35150 + }, + { + "epoch": 0.5484494914831223, + "grad_norm": 0.4142405688762665, + "learning_rate": 0.00016343774047960735, + "loss": 0.1756, + "step": 35160 + }, + { + "epoch": 0.5486054782554439, + "grad_norm": 4.836324214935303, + "learning_rate": 0.00016342734136145254, + "loss": 0.2534, + "step": 35170 + }, + { + "epoch": 0.5487614650277657, + "grad_norm": 1.4933065176010132, + "learning_rate": 0.00016341694224329779, + "loss": 0.2355, + "step": 35180 + }, + { + "epoch": 0.5489174518000873, + "grad_norm": 1.9468894004821777, + "learning_rate": 0.00016340654312514298, + "loss": 0.3113, + "step": 35190 + }, + { + "epoch": 0.5490734385724091, + "grad_norm": 1.012710452079773, + "learning_rate": 0.00016339614400698822, + "loss": 0.2635, + "step": 35200 + }, + { + "epoch": 0.5492294253447307, + "grad_norm": 1.5551140308380127, + "learning_rate": 0.00016338574488883342, + "loss": 0.2951, + "step": 35210 + }, + { + "epoch": 0.5493854121170525, + "grad_norm": 1.8613696098327637, + "learning_rate": 0.00016337534577067866, + "loss": 0.2832, + "step": 35220 + }, + { + "epoch": 0.5495413988893741, + "grad_norm": 1.6748839616775513, + "learning_rate": 0.00016336494665252386, + "loss": 0.2266, + "step": 35230 + }, + { + "epoch": 0.5496973856616959, + "grad_norm": 0.4094032347202301, + "learning_rate": 0.0001633545475343691, + "loss": 0.3497, + "step": 35240 + }, + { + "epoch": 0.5498533724340176, + "grad_norm": 1.6414631605148315, + "learning_rate": 0.0001633441484162143, + "loss": 0.1763, + "step": 35250 + }, + { + "epoch": 0.5500093592063393, + "grad_norm": 1.6380645036697388, + "learning_rate": 0.00016333374929805954, + "loss": 0.1355, + "step": 35260 + }, + { + "epoch": 0.550165345978661, + "grad_norm": 0.7325630187988281, + "learning_rate": 0.00016332335017990473, + "loss": 0.3105, + "step": 35270 + }, + { + "epoch": 0.5503213327509827, + "grad_norm": 1.6015644073486328, + "learning_rate": 0.00016331295106174998, + "loss": 0.2311, + "step": 35280 + }, + { + "epoch": 0.5504773195233045, + "grad_norm": 0.9224210381507874, + "learning_rate": 0.00016330255194359517, + "loss": 0.2746, + "step": 35290 + }, + { + "epoch": 0.5506333062956261, + "grad_norm": 1.30025315284729, + "learning_rate": 0.00016329215282544042, + "loss": 0.2066, + "step": 35300 + }, + { + "epoch": 0.5507892930679479, + "grad_norm": 0.481125146150589, + "learning_rate": 0.0001632817537072856, + "loss": 0.1778, + "step": 35310 + }, + { + "epoch": 0.5509452798402695, + "grad_norm": 0.2709486782550812, + "learning_rate": 0.00016327135458913086, + "loss": 0.1131, + "step": 35320 + }, + { + "epoch": 0.5511012666125913, + "grad_norm": 0.6248563528060913, + "learning_rate": 0.00016326095547097605, + "loss": 0.4356, + "step": 35330 + }, + { + "epoch": 0.5512572533849129, + "grad_norm": 3.0947647094726562, + "learning_rate": 0.0001632505563528213, + "loss": 0.3072, + "step": 35340 + }, + { + "epoch": 0.5514132401572347, + "grad_norm": 1.009535312652588, + "learning_rate": 0.0001632401572346665, + "loss": 0.1808, + "step": 35350 + }, + { + "epoch": 0.5515692269295563, + "grad_norm": 0.042604975402355194, + "learning_rate": 0.00016322975811651174, + "loss": 0.1198, + "step": 35360 + }, + { + "epoch": 0.5517252137018781, + "grad_norm": 1.8652396202087402, + "learning_rate": 0.00016321935899835693, + "loss": 0.2176, + "step": 35370 + }, + { + "epoch": 0.5518812004741998, + "grad_norm": 4.6097187995910645, + "learning_rate": 0.00016320895988020218, + "loss": 0.2132, + "step": 35380 + }, + { + "epoch": 0.5520371872465215, + "grad_norm": 1.3615522384643555, + "learning_rate": 0.00016319856076204737, + "loss": 0.2751, + "step": 35390 + }, + { + "epoch": 0.5521931740188432, + "grad_norm": 0.8844773769378662, + "learning_rate": 0.00016318816164389262, + "loss": 0.22, + "step": 35400 + }, + { + "epoch": 0.5523491607911649, + "grad_norm": 0.5413331985473633, + "learning_rate": 0.0001631777625257378, + "loss": 0.2159, + "step": 35410 + }, + { + "epoch": 0.5525051475634866, + "grad_norm": 1.5456678867340088, + "learning_rate": 0.00016316736340758306, + "loss": 0.2244, + "step": 35420 + }, + { + "epoch": 0.5526611343358083, + "grad_norm": 2.1405861377716064, + "learning_rate": 0.00016315696428942825, + "loss": 0.446, + "step": 35430 + }, + { + "epoch": 0.5528171211081301, + "grad_norm": 1.4269858598709106, + "learning_rate": 0.0001631465651712735, + "loss": 0.3599, + "step": 35440 + }, + { + "epoch": 0.5529731078804517, + "grad_norm": 1.8682516813278198, + "learning_rate": 0.0001631361660531187, + "loss": 0.115, + "step": 35450 + }, + { + "epoch": 0.5531290946527735, + "grad_norm": 1.0175774097442627, + "learning_rate": 0.00016312576693496394, + "loss": 0.4003, + "step": 35460 + }, + { + "epoch": 0.5532850814250951, + "grad_norm": 2.4369170665740967, + "learning_rate": 0.00016311536781680913, + "loss": 0.2489, + "step": 35470 + }, + { + "epoch": 0.5534410681974169, + "grad_norm": 1.886022686958313, + "learning_rate": 0.00016310496869865437, + "loss": 0.4201, + "step": 35480 + }, + { + "epoch": 0.5535970549697385, + "grad_norm": 1.5811959505081177, + "learning_rate": 0.00016309456958049957, + "loss": 0.4286, + "step": 35490 + }, + { + "epoch": 0.5537530417420603, + "grad_norm": 1.971110224723816, + "learning_rate": 0.00016308417046234481, + "loss": 0.3599, + "step": 35500 + }, + { + "epoch": 0.553909028514382, + "grad_norm": 0.0335380844771862, + "learning_rate": 0.00016307377134419, + "loss": 0.3357, + "step": 35510 + }, + { + "epoch": 0.5540650152867037, + "grad_norm": 1.8519577980041504, + "learning_rate": 0.00016306337222603525, + "loss": 0.3126, + "step": 35520 + }, + { + "epoch": 0.5542210020590254, + "grad_norm": 2.1563379764556885, + "learning_rate": 0.00016305297310788045, + "loss": 0.1123, + "step": 35530 + }, + { + "epoch": 0.5543769888313471, + "grad_norm": 1.64332914352417, + "learning_rate": 0.0001630425739897257, + "loss": 0.2853, + "step": 35540 + }, + { + "epoch": 0.5545329756036688, + "grad_norm": 0.061150554567575455, + "learning_rate": 0.00016303217487157088, + "loss": 0.1943, + "step": 35550 + }, + { + "epoch": 0.5546889623759905, + "grad_norm": 1.2701060771942139, + "learning_rate": 0.00016302177575341613, + "loss": 0.2304, + "step": 35560 + }, + { + "epoch": 0.5548449491483122, + "grad_norm": 2.424860715866089, + "learning_rate": 0.00016301137663526132, + "loss": 0.2128, + "step": 35570 + }, + { + "epoch": 0.5550009359206339, + "grad_norm": 0.6803575158119202, + "learning_rate": 0.00016300097751710657, + "loss": 0.1145, + "step": 35580 + }, + { + "epoch": 0.5551569226929557, + "grad_norm": 1.2855092287063599, + "learning_rate": 0.00016299057839895176, + "loss": 0.2448, + "step": 35590 + }, + { + "epoch": 0.5553129094652773, + "grad_norm": 2.6340911388397217, + "learning_rate": 0.00016298017928079698, + "loss": 0.1856, + "step": 35600 + }, + { + "epoch": 0.5554688962375991, + "grad_norm": 1.776382327079773, + "learning_rate": 0.0001629697801626422, + "loss": 0.2422, + "step": 35610 + }, + { + "epoch": 0.5556248830099207, + "grad_norm": 3.0746347904205322, + "learning_rate": 0.00016295938104448742, + "loss": 0.1578, + "step": 35620 + }, + { + "epoch": 0.5557808697822425, + "grad_norm": 1.4887659549713135, + "learning_rate": 0.00016294898192633267, + "loss": 0.2385, + "step": 35630 + }, + { + "epoch": 0.5559368565545642, + "grad_norm": 0.5371220707893372, + "learning_rate": 0.00016293858280817786, + "loss": 0.2361, + "step": 35640 + }, + { + "epoch": 0.5560928433268859, + "grad_norm": 0.18604904413223267, + "learning_rate": 0.0001629281836900231, + "loss": 0.0828, + "step": 35650 + }, + { + "epoch": 0.5562488300992076, + "grad_norm": 2.267854690551758, + "learning_rate": 0.0001629177845718683, + "loss": 0.413, + "step": 35660 + }, + { + "epoch": 0.5564048168715293, + "grad_norm": 1.9755452871322632, + "learning_rate": 0.00016290738545371355, + "loss": 0.146, + "step": 35670 + }, + { + "epoch": 0.556560803643851, + "grad_norm": 2.430293321609497, + "learning_rate": 0.00016289698633555874, + "loss": 0.2677, + "step": 35680 + }, + { + "epoch": 0.5567167904161727, + "grad_norm": 0.6319543719291687, + "learning_rate": 0.000162886587217404, + "loss": 0.1481, + "step": 35690 + }, + { + "epoch": 0.5568727771884944, + "grad_norm": 0.17439277470111847, + "learning_rate": 0.00016287618809924918, + "loss": 0.1641, + "step": 35700 + }, + { + "epoch": 0.5570287639608161, + "grad_norm": 0.350175142288208, + "learning_rate": 0.00016286578898109443, + "loss": 0.393, + "step": 35710 + }, + { + "epoch": 0.5571847507331378, + "grad_norm": 2.641941547393799, + "learning_rate": 0.00016285538986293962, + "loss": 0.1907, + "step": 35720 + }, + { + "epoch": 0.5573407375054595, + "grad_norm": 0.1910303384065628, + "learning_rate": 0.00016284499074478487, + "loss": 0.2248, + "step": 35730 + }, + { + "epoch": 0.5574967242777813, + "grad_norm": 1.0180896520614624, + "learning_rate": 0.00016283459162663006, + "loss": 0.1924, + "step": 35740 + }, + { + "epoch": 0.557652711050103, + "grad_norm": 1.2047260999679565, + "learning_rate": 0.0001628241925084753, + "loss": 0.2031, + "step": 35750 + }, + { + "epoch": 0.5578086978224247, + "grad_norm": 1.8182405233383179, + "learning_rate": 0.0001628137933903205, + "loss": 0.1573, + "step": 35760 + }, + { + "epoch": 0.5579646845947464, + "grad_norm": 0.5485963821411133, + "learning_rate": 0.00016280339427216574, + "loss": 0.152, + "step": 35770 + }, + { + "epoch": 0.5581206713670681, + "grad_norm": 0.1674145758152008, + "learning_rate": 0.00016279299515401094, + "loss": 0.2388, + "step": 35780 + }, + { + "epoch": 0.5582766581393898, + "grad_norm": 0.36187657713890076, + "learning_rate": 0.00016278259603585618, + "loss": 0.1618, + "step": 35790 + }, + { + "epoch": 0.5584326449117115, + "grad_norm": 0.013768521137535572, + "learning_rate": 0.00016277219691770138, + "loss": 0.3666, + "step": 35800 + }, + { + "epoch": 0.5585886316840332, + "grad_norm": 2.8602802753448486, + "learning_rate": 0.00016276179779954662, + "loss": 0.2065, + "step": 35810 + }, + { + "epoch": 0.5587446184563549, + "grad_norm": 1.7925455570220947, + "learning_rate": 0.00016275139868139182, + "loss": 0.1729, + "step": 35820 + }, + { + "epoch": 0.5589006052286766, + "grad_norm": 2.910456418991089, + "learning_rate": 0.00016274099956323706, + "loss": 0.3734, + "step": 35830 + }, + { + "epoch": 0.5590565920009983, + "grad_norm": 0.8378308415412903, + "learning_rate": 0.00016273060044508225, + "loss": 0.2255, + "step": 35840 + }, + { + "epoch": 0.55921257877332, + "grad_norm": 0.409534752368927, + "learning_rate": 0.0001627202013269275, + "loss": 0.2046, + "step": 35850 + }, + { + "epoch": 0.5593685655456417, + "grad_norm": 0.49498099088668823, + "learning_rate": 0.0001627098022087727, + "loss": 0.1058, + "step": 35860 + }, + { + "epoch": 0.5595245523179634, + "grad_norm": 1.1761783361434937, + "learning_rate": 0.00016269940309061794, + "loss": 0.1566, + "step": 35870 + }, + { + "epoch": 0.5596805390902851, + "grad_norm": 3.8204751014709473, + "learning_rate": 0.00016268900397246313, + "loss": 0.2646, + "step": 35880 + }, + { + "epoch": 0.5598365258626069, + "grad_norm": 0.9882522225379944, + "learning_rate": 0.00016267860485430838, + "loss": 0.1756, + "step": 35890 + }, + { + "epoch": 0.5599925126349286, + "grad_norm": 1.1832259893417358, + "learning_rate": 0.00016266820573615357, + "loss": 0.1385, + "step": 35900 + }, + { + "epoch": 0.5601484994072503, + "grad_norm": 0.7638296484947205, + "learning_rate": 0.00016265780661799882, + "loss": 0.5679, + "step": 35910 + }, + { + "epoch": 0.560304486179572, + "grad_norm": 0.8551504611968994, + "learning_rate": 0.000162647407499844, + "loss": 0.2981, + "step": 35920 + }, + { + "epoch": 0.5604604729518937, + "grad_norm": 5.499948501586914, + "learning_rate": 0.00016263700838168926, + "loss": 0.1429, + "step": 35930 + }, + { + "epoch": 0.5606164597242154, + "grad_norm": 3.0494496822357178, + "learning_rate": 0.00016262660926353445, + "loss": 0.1322, + "step": 35940 + }, + { + "epoch": 0.5607724464965371, + "grad_norm": 0.2482384592294693, + "learning_rate": 0.0001626162101453797, + "loss": 0.2473, + "step": 35950 + }, + { + "epoch": 0.5609284332688588, + "grad_norm": 0.009750776924192905, + "learning_rate": 0.0001626058110272249, + "loss": 0.0429, + "step": 35960 + }, + { + "epoch": 0.5610844200411805, + "grad_norm": 0.5742604732513428, + "learning_rate": 0.00016259541190907014, + "loss": 0.4565, + "step": 35970 + }, + { + "epoch": 0.5612404068135022, + "grad_norm": 0.7810243368148804, + "learning_rate": 0.00016258501279091533, + "loss": 0.2107, + "step": 35980 + }, + { + "epoch": 0.5613963935858239, + "grad_norm": 2.785747528076172, + "learning_rate": 0.00016257461367276058, + "loss": 0.5087, + "step": 35990 + }, + { + "epoch": 0.5615523803581456, + "grad_norm": 0.8806902766227722, + "learning_rate": 0.00016256421455460577, + "loss": 0.2937, + "step": 36000 + }, + { + "epoch": 0.5617083671304673, + "grad_norm": 1.857373833656311, + "learning_rate": 0.00016255381543645102, + "loss": 0.5042, + "step": 36010 + }, + { + "epoch": 0.561864353902789, + "grad_norm": 0.5333901047706604, + "learning_rate": 0.0001625434163182962, + "loss": 0.1678, + "step": 36020 + }, + { + "epoch": 0.5620203406751108, + "grad_norm": 11.023160934448242, + "learning_rate": 0.00016253301720014145, + "loss": 0.2721, + "step": 36030 + }, + { + "epoch": 0.5621763274474325, + "grad_norm": 2.6214029788970947, + "learning_rate": 0.00016252261808198665, + "loss": 0.1675, + "step": 36040 + }, + { + "epoch": 0.5623323142197542, + "grad_norm": 0.25980064272880554, + "learning_rate": 0.0001625122189638319, + "loss": 0.1832, + "step": 36050 + }, + { + "epoch": 0.5624883009920759, + "grad_norm": 1.3559473752975464, + "learning_rate": 0.0001625018198456771, + "loss": 0.121, + "step": 36060 + }, + { + "epoch": 0.5626442877643976, + "grad_norm": 0.3558153212070465, + "learning_rate": 0.0001624914207275223, + "loss": 0.2797, + "step": 36070 + }, + { + "epoch": 0.5628002745367193, + "grad_norm": 0.6314427256584167, + "learning_rate": 0.00016248102160936753, + "loss": 0.3455, + "step": 36080 + }, + { + "epoch": 0.562956261309041, + "grad_norm": 0.16120101511478424, + "learning_rate": 0.00016247062249121275, + "loss": 0.3669, + "step": 36090 + }, + { + "epoch": 0.5631122480813627, + "grad_norm": 1.041572093963623, + "learning_rate": 0.00016246022337305797, + "loss": 0.2744, + "step": 36100 + }, + { + "epoch": 0.5632682348536844, + "grad_norm": 0.2645890414714813, + "learning_rate": 0.00016244982425490318, + "loss": 0.1487, + "step": 36110 + }, + { + "epoch": 0.5634242216260061, + "grad_norm": 5.698698043823242, + "learning_rate": 0.0001624394251367484, + "loss": 0.173, + "step": 36120 + }, + { + "epoch": 0.5635802083983278, + "grad_norm": 0.3299804627895355, + "learning_rate": 0.00016242902601859362, + "loss": 0.1965, + "step": 36130 + }, + { + "epoch": 0.5637361951706495, + "grad_norm": 1.2476481199264526, + "learning_rate": 0.00016241862690043884, + "loss": 0.4097, + "step": 36140 + }, + { + "epoch": 0.5638921819429712, + "grad_norm": 0.30926263332366943, + "learning_rate": 0.00016240822778228406, + "loss": 0.1416, + "step": 36150 + }, + { + "epoch": 0.564048168715293, + "grad_norm": 0.05872740224003792, + "learning_rate": 0.00016239782866412928, + "loss": 0.1449, + "step": 36160 + }, + { + "epoch": 0.5642041554876146, + "grad_norm": 0.2292211800813675, + "learning_rate": 0.0001623874295459745, + "loss": 0.2557, + "step": 36170 + }, + { + "epoch": 0.5643601422599364, + "grad_norm": 1.7822531461715698, + "learning_rate": 0.00016237703042781972, + "loss": 0.2953, + "step": 36180 + }, + { + "epoch": 0.5645161290322581, + "grad_norm": 2.4908461570739746, + "learning_rate": 0.00016236663130966494, + "loss": 0.3888, + "step": 36190 + }, + { + "epoch": 0.5646721158045798, + "grad_norm": 0.08363594114780426, + "learning_rate": 0.00016235623219151016, + "loss": 0.2293, + "step": 36200 + }, + { + "epoch": 0.5648281025769015, + "grad_norm": 1.4830002784729004, + "learning_rate": 0.00016234583307335538, + "loss": 0.2348, + "step": 36210 + }, + { + "epoch": 0.5649840893492232, + "grad_norm": 1.8443071842193604, + "learning_rate": 0.0001623354339552006, + "loss": 0.3392, + "step": 36220 + }, + { + "epoch": 0.5651400761215449, + "grad_norm": 2.8050875663757324, + "learning_rate": 0.00016232503483704582, + "loss": 0.2426, + "step": 36230 + }, + { + "epoch": 0.5652960628938666, + "grad_norm": 3.6627259254455566, + "learning_rate": 0.00016231463571889104, + "loss": 0.2118, + "step": 36240 + }, + { + "epoch": 0.5654520496661883, + "grad_norm": 0.3443094789981842, + "learning_rate": 0.00016230423660073626, + "loss": 0.2418, + "step": 36250 + }, + { + "epoch": 0.56560803643851, + "grad_norm": 1.7411112785339355, + "learning_rate": 0.00016229383748258148, + "loss": 0.1628, + "step": 36260 + }, + { + "epoch": 0.5657640232108317, + "grad_norm": 1.1321816444396973, + "learning_rate": 0.0001622834383644267, + "loss": 0.2732, + "step": 36270 + }, + { + "epoch": 0.5659200099831534, + "grad_norm": 2.9686615467071533, + "learning_rate": 0.00016227303924627192, + "loss": 0.3036, + "step": 36280 + }, + { + "epoch": 0.5660759967554752, + "grad_norm": 5.03530740737915, + "learning_rate": 0.00016226264012811714, + "loss": 0.2143, + "step": 36290 + }, + { + "epoch": 0.5662319835277968, + "grad_norm": 2.3943281173706055, + "learning_rate": 0.00016225224100996236, + "loss": 0.2629, + "step": 36300 + }, + { + "epoch": 0.5663879703001186, + "grad_norm": 1.812828779220581, + "learning_rate": 0.00016224184189180758, + "loss": 0.3892, + "step": 36310 + }, + { + "epoch": 0.5665439570724402, + "grad_norm": 1.4708483219146729, + "learning_rate": 0.0001622314427736528, + "loss": 0.2118, + "step": 36320 + }, + { + "epoch": 0.566699943844762, + "grad_norm": 8.99713134765625, + "learning_rate": 0.00016222104365549802, + "loss": 0.3185, + "step": 36330 + }, + { + "epoch": 0.5668559306170837, + "grad_norm": 1.7472341060638428, + "learning_rate": 0.00016221064453734324, + "loss": 0.1659, + "step": 36340 + }, + { + "epoch": 0.5670119173894054, + "grad_norm": 0.8059778213500977, + "learning_rate": 0.00016220024541918846, + "loss": 0.0621, + "step": 36350 + }, + { + "epoch": 0.5671679041617271, + "grad_norm": 4.0501861572265625, + "learning_rate": 0.00016218984630103368, + "loss": 0.1399, + "step": 36360 + }, + { + "epoch": 0.5673238909340488, + "grad_norm": 0.014400321058928967, + "learning_rate": 0.0001621794471828789, + "loss": 0.2372, + "step": 36370 + }, + { + "epoch": 0.5674798777063705, + "grad_norm": 0.4061933755874634, + "learning_rate": 0.00016216904806472412, + "loss": 0.3854, + "step": 36380 + }, + { + "epoch": 0.5676358644786922, + "grad_norm": 0.4940955340862274, + "learning_rate": 0.00016215864894656933, + "loss": 0.1689, + "step": 36390 + }, + { + "epoch": 0.567791851251014, + "grad_norm": 0.3220517635345459, + "learning_rate": 0.00016214824982841455, + "loss": 0.1912, + "step": 36400 + }, + { + "epoch": 0.5679478380233356, + "grad_norm": 2.0848920345306396, + "learning_rate": 0.00016213785071025977, + "loss": 0.1639, + "step": 36410 + }, + { + "epoch": 0.5681038247956574, + "grad_norm": 0.04129406809806824, + "learning_rate": 0.000162127451592105, + "loss": 0.1378, + "step": 36420 + }, + { + "epoch": 0.568259811567979, + "grad_norm": 1.9343401193618774, + "learning_rate": 0.0001621170524739502, + "loss": 0.3759, + "step": 36430 + }, + { + "epoch": 0.5684157983403008, + "grad_norm": 1.0625615119934082, + "learning_rate": 0.00016210665335579543, + "loss": 0.2074, + "step": 36440 + }, + { + "epoch": 0.5685717851126224, + "grad_norm": 0.14784491062164307, + "learning_rate": 0.00016209625423764065, + "loss": 0.0981, + "step": 36450 + }, + { + "epoch": 0.5687277718849442, + "grad_norm": 1.8117821216583252, + "learning_rate": 0.00016208585511948587, + "loss": 0.2755, + "step": 36460 + }, + { + "epoch": 0.5688837586572658, + "grad_norm": 0.061010006815195084, + "learning_rate": 0.0001620754560013311, + "loss": 0.1815, + "step": 36470 + }, + { + "epoch": 0.5690397454295876, + "grad_norm": 0.44453224539756775, + "learning_rate": 0.0001620650568831763, + "loss": 0.0891, + "step": 36480 + }, + { + "epoch": 0.5691957322019093, + "grad_norm": 0.6130645871162415, + "learning_rate": 0.00016205465776502153, + "loss": 0.2778, + "step": 36490 + }, + { + "epoch": 0.569351718974231, + "grad_norm": 5.408796787261963, + "learning_rate": 0.00016204425864686678, + "loss": 0.2301, + "step": 36500 + }, + { + "epoch": 0.5695077057465527, + "grad_norm": 1.2095412015914917, + "learning_rate": 0.00016203385952871197, + "loss": 0.1378, + "step": 36510 + }, + { + "epoch": 0.5696636925188744, + "grad_norm": 0.46873530745506287, + "learning_rate": 0.0001620234604105572, + "loss": 0.1431, + "step": 36520 + }, + { + "epoch": 0.5698196792911961, + "grad_norm": 1.028438925743103, + "learning_rate": 0.0001620130612924024, + "loss": 0.1114, + "step": 36530 + }, + { + "epoch": 0.5699756660635178, + "grad_norm": 0.5952587127685547, + "learning_rate": 0.00016200266217424763, + "loss": 0.2156, + "step": 36540 + }, + { + "epoch": 0.5701316528358396, + "grad_norm": 0.5134285688400269, + "learning_rate": 0.00016199226305609285, + "loss": 0.2712, + "step": 36550 + }, + { + "epoch": 0.5702876396081612, + "grad_norm": 0.03440163657069206, + "learning_rate": 0.00016198186393793807, + "loss": 0.312, + "step": 36560 + }, + { + "epoch": 0.570443626380483, + "grad_norm": 0.10363951325416565, + "learning_rate": 0.0001619714648197833, + "loss": 0.2399, + "step": 36570 + }, + { + "epoch": 0.5705996131528046, + "grad_norm": 1.3341947793960571, + "learning_rate": 0.0001619610657016285, + "loss": 0.1397, + "step": 36580 + }, + { + "epoch": 0.5707555999251264, + "grad_norm": 0.0668073296546936, + "learning_rate": 0.00016195066658347373, + "loss": 0.1091, + "step": 36590 + }, + { + "epoch": 0.570911586697448, + "grad_norm": 5.0486674308776855, + "learning_rate": 0.00016194026746531895, + "loss": 0.4893, + "step": 36600 + }, + { + "epoch": 0.5710675734697698, + "grad_norm": 1.2655692100524902, + "learning_rate": 0.00016192986834716417, + "loss": 0.3344, + "step": 36610 + }, + { + "epoch": 0.5712235602420914, + "grad_norm": 0.1458190381526947, + "learning_rate": 0.00016191946922900939, + "loss": 0.247, + "step": 36620 + }, + { + "epoch": 0.5713795470144132, + "grad_norm": 2.5789072513580322, + "learning_rate": 0.0001619090701108546, + "loss": 0.1546, + "step": 36630 + }, + { + "epoch": 0.5715355337867349, + "grad_norm": 1.191821575164795, + "learning_rate": 0.00016189867099269983, + "loss": 0.2215, + "step": 36640 + }, + { + "epoch": 0.5716915205590566, + "grad_norm": 8.163617134094238, + "learning_rate": 0.00016188827187454505, + "loss": 0.1696, + "step": 36650 + }, + { + "epoch": 0.5718475073313783, + "grad_norm": 3.162277936935425, + "learning_rate": 0.00016187787275639027, + "loss": 0.2906, + "step": 36660 + }, + { + "epoch": 0.5720034941037, + "grad_norm": 2.2368602752685547, + "learning_rate": 0.00016186747363823548, + "loss": 0.3806, + "step": 36670 + }, + { + "epoch": 0.5721594808760218, + "grad_norm": 2.0193803310394287, + "learning_rate": 0.0001618570745200807, + "loss": 0.1964, + "step": 36680 + }, + { + "epoch": 0.5723154676483434, + "grad_norm": 0.9131811261177063, + "learning_rate": 0.00016184667540192592, + "loss": 0.1902, + "step": 36690 + }, + { + "epoch": 0.5724714544206652, + "grad_norm": 0.21330870687961578, + "learning_rate": 0.00016183627628377114, + "loss": 0.2907, + "step": 36700 + }, + { + "epoch": 0.5726274411929868, + "grad_norm": 0.7145895957946777, + "learning_rate": 0.00016182587716561636, + "loss": 0.2574, + "step": 36710 + }, + { + "epoch": 0.5727834279653086, + "grad_norm": 1.0699433088302612, + "learning_rate": 0.00016181547804746158, + "loss": 0.1675, + "step": 36720 + }, + { + "epoch": 0.5729394147376302, + "grad_norm": 0.5066679120063782, + "learning_rate": 0.0001618050789293068, + "loss": 0.4487, + "step": 36730 + }, + { + "epoch": 0.573095401509952, + "grad_norm": 0.7482333183288574, + "learning_rate": 0.00016179467981115202, + "loss": 0.4101, + "step": 36740 + }, + { + "epoch": 0.5732513882822736, + "grad_norm": 0.15925532579421997, + "learning_rate": 0.00016178428069299724, + "loss": 0.2825, + "step": 36750 + }, + { + "epoch": 0.5734073750545954, + "grad_norm": 0.95689457654953, + "learning_rate": 0.00016177388157484246, + "loss": 0.2754, + "step": 36760 + }, + { + "epoch": 0.573563361826917, + "grad_norm": 2.15043568611145, + "learning_rate": 0.00016176348245668768, + "loss": 0.2073, + "step": 36770 + }, + { + "epoch": 0.5737193485992388, + "grad_norm": 2.651520252227783, + "learning_rate": 0.0001617530833385329, + "loss": 0.2517, + "step": 36780 + }, + { + "epoch": 0.5738753353715605, + "grad_norm": 2.1156702041625977, + "learning_rate": 0.00016174268422037812, + "loss": 0.3029, + "step": 36790 + }, + { + "epoch": 0.5740313221438822, + "grad_norm": 3.6200709342956543, + "learning_rate": 0.00016173228510222334, + "loss": 0.4174, + "step": 36800 + }, + { + "epoch": 0.574187308916204, + "grad_norm": 1.447937250137329, + "learning_rate": 0.00016172188598406856, + "loss": 0.1763, + "step": 36810 + }, + { + "epoch": 0.5743432956885256, + "grad_norm": 1.4020819664001465, + "learning_rate": 0.00016171148686591378, + "loss": 0.3994, + "step": 36820 + }, + { + "epoch": 0.5744992824608474, + "grad_norm": 1.4079807996749878, + "learning_rate": 0.000161701087747759, + "loss": 0.2051, + "step": 36830 + }, + { + "epoch": 0.574655269233169, + "grad_norm": 0.05268567427992821, + "learning_rate": 0.00016169068862960422, + "loss": 0.1773, + "step": 36840 + }, + { + "epoch": 0.5748112560054908, + "grad_norm": 2.248767614364624, + "learning_rate": 0.00016168028951144944, + "loss": 0.2534, + "step": 36850 + }, + { + "epoch": 0.5749672427778124, + "grad_norm": 0.3726302981376648, + "learning_rate": 0.00016166989039329466, + "loss": 0.328, + "step": 36860 + }, + { + "epoch": 0.5751232295501342, + "grad_norm": 1.0772110223770142, + "learning_rate": 0.00016165949127513988, + "loss": 0.2069, + "step": 36870 + }, + { + "epoch": 0.5752792163224558, + "grad_norm": 0.11339398473501205, + "learning_rate": 0.0001616490921569851, + "loss": 0.1827, + "step": 36880 + }, + { + "epoch": 0.5754352030947776, + "grad_norm": 1.2981103658676147, + "learning_rate": 0.00016163869303883032, + "loss": 0.1378, + "step": 36890 + }, + { + "epoch": 0.5755911898670992, + "grad_norm": 0.3302285075187683, + "learning_rate": 0.00016162829392067554, + "loss": 0.2311, + "step": 36900 + }, + { + "epoch": 0.575747176639421, + "grad_norm": 0.10168436169624329, + "learning_rate": 0.00016161789480252076, + "loss": 0.1678, + "step": 36910 + }, + { + "epoch": 0.5759031634117426, + "grad_norm": 1.6778357028961182, + "learning_rate": 0.00016160749568436598, + "loss": 0.2385, + "step": 36920 + }, + { + "epoch": 0.5760591501840644, + "grad_norm": 8.1710205078125, + "learning_rate": 0.0001615970965662112, + "loss": 0.3287, + "step": 36930 + }, + { + "epoch": 0.5762151369563862, + "grad_norm": 2.6481053829193115, + "learning_rate": 0.00016158669744805641, + "loss": 0.3119, + "step": 36940 + }, + { + "epoch": 0.5763711237287078, + "grad_norm": 0.3120105266571045, + "learning_rate": 0.00016157629832990163, + "loss": 0.1985, + "step": 36950 + }, + { + "epoch": 0.5765271105010296, + "grad_norm": 2.7221410274505615, + "learning_rate": 0.00016156589921174685, + "loss": 0.3697, + "step": 36960 + }, + { + "epoch": 0.5766830972733512, + "grad_norm": 0.34397706389427185, + "learning_rate": 0.00016155550009359207, + "loss": 0.2162, + "step": 36970 + }, + { + "epoch": 0.576839084045673, + "grad_norm": 4.6421003341674805, + "learning_rate": 0.0001615451009754373, + "loss": 0.2892, + "step": 36980 + }, + { + "epoch": 0.5769950708179946, + "grad_norm": 1.5395855903625488, + "learning_rate": 0.0001615347018572825, + "loss": 0.1258, + "step": 36990 + }, + { + "epoch": 0.5771510575903164, + "grad_norm": 0.4586782455444336, + "learning_rate": 0.00016152430273912773, + "loss": 0.1752, + "step": 37000 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.253144195072e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..663bf108e06e6e6304a43fbab9e53a3fd7099579 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8856339147eabbe1be6d108defc08e47b7924451dd12fd9743990bb1364a08c1 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..4471dc3476d80797b7b3ebe47598abb1c7ebedb6 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1b70256d46f60d17ccfac5cabe0886d321ab61498c9e7f30f4d053fb44064d6 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..0c350271682568c777035f229e4da3fd82f4ed24 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..36b9286de055f39f2420a15794276cdfdfda7ba0 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..e953a9bf018bf94fb1ff4eeb94e1d24ad5aa7b41 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/trainer_state.json @@ -0,0 +1,26284 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5849503962064017, + "eval_steps": 500, + "global_step": 37500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + }, + { + "epoch": 0.5071129968178698, + "grad_norm": 1.0075199604034424, + "learning_rate": 0.00016619350679062414, + "loss": 0.2569, + "step": 32510 + }, + { + "epoch": 0.5072689835901916, + "grad_norm": 1.3489066362380981, + "learning_rate": 0.0001661831076724694, + "loss": 0.2565, + "step": 32520 + }, + { + "epoch": 0.5074249703625132, + "grad_norm": 0.2534058094024658, + "learning_rate": 0.00016617270855431458, + "loss": 0.2292, + "step": 32530 + }, + { + "epoch": 0.507580957134835, + "grad_norm": 0.31723281741142273, + "learning_rate": 0.00016616230943615983, + "loss": 0.1467, + "step": 32540 + }, + { + "epoch": 0.5077369439071566, + "grad_norm": 0.29659849405288696, + "learning_rate": 0.00016615191031800502, + "loss": 0.2768, + "step": 32550 + }, + { + "epoch": 0.5078929306794784, + "grad_norm": 1.5246435403823853, + "learning_rate": 0.00016614151119985027, + "loss": 0.216, + "step": 32560 + }, + { + "epoch": 0.5080489174518001, + "grad_norm": 0.6026854515075684, + "learning_rate": 0.00016613111208169546, + "loss": 0.2277, + "step": 32570 + }, + { + "epoch": 0.5082049042241218, + "grad_norm": 0.41994890570640564, + "learning_rate": 0.0001661207129635407, + "loss": 0.2433, + "step": 32580 + }, + { + "epoch": 0.5083608909964435, + "grad_norm": 2.084599733352661, + "learning_rate": 0.00016611031384538593, + "loss": 0.4081, + "step": 32590 + }, + { + "epoch": 0.5085168777687652, + "grad_norm": 0.2945229709148407, + "learning_rate": 0.00016609991472723115, + "loss": 0.1329, + "step": 32600 + }, + { + "epoch": 0.508672864541087, + "grad_norm": 1.4981423616409302, + "learning_rate": 0.00016608951560907636, + "loss": 0.2855, + "step": 32610 + }, + { + "epoch": 0.5088288513134086, + "grad_norm": 0.961737871170044, + "learning_rate": 0.00016607911649092158, + "loss": 0.1654, + "step": 32620 + }, + { + "epoch": 0.5089848380857304, + "grad_norm": 1.49734628200531, + "learning_rate": 0.0001660687173727668, + "loss": 0.2081, + "step": 32630 + }, + { + "epoch": 0.509140824858052, + "grad_norm": 2.538135051727295, + "learning_rate": 0.00016605831825461202, + "loss": 0.3023, + "step": 32640 + }, + { + "epoch": 0.5092968116303738, + "grad_norm": 0.2981272041797638, + "learning_rate": 0.00016604791913645724, + "loss": 0.214, + "step": 32650 + }, + { + "epoch": 0.5094527984026954, + "grad_norm": 1.8863650560379028, + "learning_rate": 0.00016603752001830246, + "loss": 0.3363, + "step": 32660 + }, + { + "epoch": 0.5096087851750172, + "grad_norm": 0.022720983251929283, + "learning_rate": 0.00016602712090014768, + "loss": 0.2463, + "step": 32670 + }, + { + "epoch": 0.5097647719473388, + "grad_norm": 1.154977798461914, + "learning_rate": 0.0001660167217819929, + "loss": 0.4484, + "step": 32680 + }, + { + "epoch": 0.5099207587196606, + "grad_norm": 0.9470499157905579, + "learning_rate": 0.00016600632266383812, + "loss": 0.1196, + "step": 32690 + }, + { + "epoch": 0.5100767454919822, + "grad_norm": 0.19955401122570038, + "learning_rate": 0.00016599592354568334, + "loss": 0.1762, + "step": 32700 + }, + { + "epoch": 0.510232732264304, + "grad_norm": 0.4098282754421234, + "learning_rate": 0.00016598552442752856, + "loss": 0.2739, + "step": 32710 + }, + { + "epoch": 0.5103887190366257, + "grad_norm": 0.21697258949279785, + "learning_rate": 0.00016597512530937378, + "loss": 0.2555, + "step": 32720 + }, + { + "epoch": 0.5105447058089474, + "grad_norm": 4.894479751586914, + "learning_rate": 0.000165964726191219, + "loss": 0.3137, + "step": 32730 + }, + { + "epoch": 0.5107006925812692, + "grad_norm": 0.8396711349487305, + "learning_rate": 0.00016595432707306422, + "loss": 0.3105, + "step": 32740 + }, + { + "epoch": 0.5108566793535908, + "grad_norm": 1.5494674444198608, + "learning_rate": 0.00016594392795490944, + "loss": 0.2271, + "step": 32750 + }, + { + "epoch": 0.5110126661259126, + "grad_norm": 1.3409888744354248, + "learning_rate": 0.00016593352883675466, + "loss": 0.1796, + "step": 32760 + }, + { + "epoch": 0.5111686528982342, + "grad_norm": 1.7590911388397217, + "learning_rate": 0.00016592312971859988, + "loss": 0.2185, + "step": 32770 + }, + { + "epoch": 0.511324639670556, + "grad_norm": 0.7350002527236938, + "learning_rate": 0.0001659127306004451, + "loss": 0.1911, + "step": 32780 + }, + { + "epoch": 0.5114806264428776, + "grad_norm": 0.7176411747932434, + "learning_rate": 0.00016590233148229032, + "loss": 0.6918, + "step": 32790 + }, + { + "epoch": 0.5116366132151994, + "grad_norm": 2.5005321502685547, + "learning_rate": 0.00016589193236413554, + "loss": 0.2329, + "step": 32800 + }, + { + "epoch": 0.511792599987521, + "grad_norm": 1.3916250467300415, + "learning_rate": 0.00016588153324598076, + "loss": 0.2102, + "step": 32810 + }, + { + "epoch": 0.5119485867598428, + "grad_norm": 0.721671998500824, + "learning_rate": 0.00016587113412782598, + "loss": 0.2965, + "step": 32820 + }, + { + "epoch": 0.5121045735321644, + "grad_norm": 0.6120114326477051, + "learning_rate": 0.0001658607350096712, + "loss": 0.2661, + "step": 32830 + }, + { + "epoch": 0.5122605603044862, + "grad_norm": 0.29123347997665405, + "learning_rate": 0.00016585033589151642, + "loss": 0.2455, + "step": 32840 + }, + { + "epoch": 0.5124165470768078, + "grad_norm": 2.7300868034362793, + "learning_rate": 0.00016583993677336164, + "loss": 0.2049, + "step": 32850 + }, + { + "epoch": 0.5125725338491296, + "grad_norm": 1.801928997039795, + "learning_rate": 0.00016582953765520686, + "loss": 0.1725, + "step": 32860 + }, + { + "epoch": 0.5127285206214514, + "grad_norm": 0.6406471729278564, + "learning_rate": 0.00016581913853705208, + "loss": 0.164, + "step": 32870 + }, + { + "epoch": 0.512884507393773, + "grad_norm": 0.9627325534820557, + "learning_rate": 0.0001658087394188973, + "loss": 0.1728, + "step": 32880 + }, + { + "epoch": 0.5130404941660948, + "grad_norm": 1.5446962118148804, + "learning_rate": 0.00016579834030074251, + "loss": 0.2345, + "step": 32890 + }, + { + "epoch": 0.5131964809384164, + "grad_norm": 0.28092706203460693, + "learning_rate": 0.0001657879411825877, + "loss": 0.2124, + "step": 32900 + }, + { + "epoch": 0.5133524677107382, + "grad_norm": 1.7751421928405762, + "learning_rate": 0.00016577754206443295, + "loss": 0.2367, + "step": 32910 + }, + { + "epoch": 0.5135084544830598, + "grad_norm": 1.6964634656906128, + "learning_rate": 0.00016576714294627815, + "loss": 0.2361, + "step": 32920 + }, + { + "epoch": 0.5136644412553816, + "grad_norm": 0.06759823113679886, + "learning_rate": 0.0001657567438281234, + "loss": 0.2329, + "step": 32930 + }, + { + "epoch": 0.5138204280277032, + "grad_norm": 2.2519280910491943, + "learning_rate": 0.00016574634470996859, + "loss": 0.1079, + "step": 32940 + }, + { + "epoch": 0.513976414800025, + "grad_norm": 3.167576551437378, + "learning_rate": 0.00016573594559181383, + "loss": 0.1764, + "step": 32950 + }, + { + "epoch": 0.5141324015723466, + "grad_norm": 1.6734397411346436, + "learning_rate": 0.00016572554647365903, + "loss": 0.2878, + "step": 32960 + }, + { + "epoch": 0.5142883883446684, + "grad_norm": 0.35627374053001404, + "learning_rate": 0.00016571514735550427, + "loss": 0.2337, + "step": 32970 + }, + { + "epoch": 0.51444437511699, + "grad_norm": 1.6321039199829102, + "learning_rate": 0.00016570474823734946, + "loss": 0.2551, + "step": 32980 + }, + { + "epoch": 0.5146003618893118, + "grad_norm": 1.352982997894287, + "learning_rate": 0.0001656943491191947, + "loss": 0.4467, + "step": 32990 + }, + { + "epoch": 0.5147563486616334, + "grad_norm": 0.3558129072189331, + "learning_rate": 0.0001656839500010399, + "loss": 0.1913, + "step": 33000 + }, + { + "epoch": 0.5149123354339552, + "grad_norm": 0.041972286999225616, + "learning_rate": 0.00016567355088288515, + "loss": 0.1704, + "step": 33010 + }, + { + "epoch": 0.515068322206277, + "grad_norm": 1.4168939590454102, + "learning_rate": 0.00016566315176473034, + "loss": 0.4866, + "step": 33020 + }, + { + "epoch": 0.5152243089785986, + "grad_norm": 3.0042929649353027, + "learning_rate": 0.0001656527526465756, + "loss": 0.2473, + "step": 33030 + }, + { + "epoch": 0.5153802957509204, + "grad_norm": 1.027274489402771, + "learning_rate": 0.00016564235352842078, + "loss": 0.436, + "step": 33040 + }, + { + "epoch": 0.515536282523242, + "grad_norm": 1.2330195903778076, + "learning_rate": 0.00016563195441026603, + "loss": 0.4615, + "step": 33050 + }, + { + "epoch": 0.5156922692955638, + "grad_norm": 2.7145066261291504, + "learning_rate": 0.00016562155529211122, + "loss": 0.278, + "step": 33060 + }, + { + "epoch": 0.5158482560678854, + "grad_norm": 1.1273893117904663, + "learning_rate": 0.00016561115617395647, + "loss": 0.3767, + "step": 33070 + }, + { + "epoch": 0.5160042428402072, + "grad_norm": 2.466265916824341, + "learning_rate": 0.00016560075705580166, + "loss": 0.1879, + "step": 33080 + }, + { + "epoch": 0.5161602296125288, + "grad_norm": 0.3759515583515167, + "learning_rate": 0.0001655903579376469, + "loss": 0.4004, + "step": 33090 + }, + { + "epoch": 0.5163162163848506, + "grad_norm": 1.8821336030960083, + "learning_rate": 0.0001655799588194921, + "loss": 0.1622, + "step": 33100 + }, + { + "epoch": 0.5164722031571722, + "grad_norm": 0.8162096738815308, + "learning_rate": 0.00016556955970133735, + "loss": 0.1341, + "step": 33110 + }, + { + "epoch": 0.516628189929494, + "grad_norm": 0.008498801849782467, + "learning_rate": 0.00016555916058318254, + "loss": 0.2546, + "step": 33120 + }, + { + "epoch": 0.5167841767018156, + "grad_norm": 0.17536675930023193, + "learning_rate": 0.00016554876146502779, + "loss": 0.2934, + "step": 33130 + }, + { + "epoch": 0.5169401634741374, + "grad_norm": 3.539898633956909, + "learning_rate": 0.00016553836234687298, + "loss": 0.4959, + "step": 33140 + }, + { + "epoch": 0.517096150246459, + "grad_norm": 2.190887212753296, + "learning_rate": 0.00016552796322871823, + "loss": 0.2914, + "step": 33150 + }, + { + "epoch": 0.5172521370187808, + "grad_norm": 1.6388943195343018, + "learning_rate": 0.00016551756411056342, + "loss": 0.1317, + "step": 33160 + }, + { + "epoch": 0.5174081237911026, + "grad_norm": 0.33063387870788574, + "learning_rate": 0.00016550716499240866, + "loss": 0.4258, + "step": 33170 + }, + { + "epoch": 0.5175641105634242, + "grad_norm": 0.5168288946151733, + "learning_rate": 0.00016549676587425386, + "loss": 0.2548, + "step": 33180 + }, + { + "epoch": 0.517720097335746, + "grad_norm": 3.9570281505584717, + "learning_rate": 0.0001654863667560991, + "loss": 0.3179, + "step": 33190 + }, + { + "epoch": 0.5178760841080676, + "grad_norm": 1.127255916595459, + "learning_rate": 0.0001654759676379443, + "loss": 0.233, + "step": 33200 + }, + { + "epoch": 0.5180320708803894, + "grad_norm": 0.36299410462379456, + "learning_rate": 0.00016546556851978954, + "loss": 0.4614, + "step": 33210 + }, + { + "epoch": 0.518188057652711, + "grad_norm": 2.914668321609497, + "learning_rate": 0.00016545516940163474, + "loss": 0.1603, + "step": 33220 + }, + { + "epoch": 0.5183440444250328, + "grad_norm": 6.991094589233398, + "learning_rate": 0.00016544477028347998, + "loss": 0.2259, + "step": 33230 + }, + { + "epoch": 0.5185000311973544, + "grad_norm": 2.0594241619110107, + "learning_rate": 0.00016543437116532517, + "loss": 0.1885, + "step": 33240 + }, + { + "epoch": 0.5186560179696762, + "grad_norm": 0.5594693422317505, + "learning_rate": 0.00016542397204717042, + "loss": 0.4333, + "step": 33250 + }, + { + "epoch": 0.5188120047419978, + "grad_norm": 4.314854621887207, + "learning_rate": 0.00016541357292901561, + "loss": 0.2103, + "step": 33260 + }, + { + "epoch": 0.5189679915143196, + "grad_norm": 0.5609000325202942, + "learning_rate": 0.00016540317381086086, + "loss": 0.2498, + "step": 33270 + }, + { + "epoch": 0.5191239782866413, + "grad_norm": 3.680088758468628, + "learning_rate": 0.00016539277469270605, + "loss": 0.3465, + "step": 33280 + }, + { + "epoch": 0.519279965058963, + "grad_norm": 1.4993672370910645, + "learning_rate": 0.0001653823755745513, + "loss": 0.3272, + "step": 33290 + }, + { + "epoch": 0.5194359518312847, + "grad_norm": 0.18005084991455078, + "learning_rate": 0.0001653719764563965, + "loss": 0.1706, + "step": 33300 + }, + { + "epoch": 0.5195919386036064, + "grad_norm": 0.5105524659156799, + "learning_rate": 0.00016536157733824174, + "loss": 0.0937, + "step": 33310 + }, + { + "epoch": 0.5197479253759281, + "grad_norm": 1.2806589603424072, + "learning_rate": 0.00016535117822008693, + "loss": 0.1578, + "step": 33320 + }, + { + "epoch": 0.5199039121482498, + "grad_norm": 0.22015990316867828, + "learning_rate": 0.00016534077910193218, + "loss": 0.2081, + "step": 33330 + }, + { + "epoch": 0.5200598989205716, + "grad_norm": 0.34638991951942444, + "learning_rate": 0.00016533037998377737, + "loss": 0.2389, + "step": 33340 + }, + { + "epoch": 0.5202158856928932, + "grad_norm": 0.8709302544593811, + "learning_rate": 0.0001653199808656226, + "loss": 0.2112, + "step": 33350 + }, + { + "epoch": 0.520371872465215, + "grad_norm": 1.2309521436691284, + "learning_rate": 0.0001653095817474678, + "loss": 0.3448, + "step": 33360 + }, + { + "epoch": 0.5205278592375366, + "grad_norm": 1.3155115842819214, + "learning_rate": 0.00016529918262931303, + "loss": 0.1572, + "step": 33370 + }, + { + "epoch": 0.5206838460098584, + "grad_norm": 0.358751118183136, + "learning_rate": 0.00016528878351115825, + "loss": 0.3011, + "step": 33380 + }, + { + "epoch": 0.52083983278218, + "grad_norm": 3.800628662109375, + "learning_rate": 0.00016527838439300347, + "loss": 0.458, + "step": 33390 + }, + { + "epoch": 0.5209958195545018, + "grad_norm": 1.1253124475479126, + "learning_rate": 0.0001652679852748487, + "loss": 0.1238, + "step": 33400 + }, + { + "epoch": 0.5211518063268235, + "grad_norm": 0.46979108452796936, + "learning_rate": 0.0001652575861566939, + "loss": 0.0745, + "step": 33410 + }, + { + "epoch": 0.5213077930991452, + "grad_norm": 7.921492576599121, + "learning_rate": 0.00016524718703853913, + "loss": 0.4004, + "step": 33420 + }, + { + "epoch": 0.5214637798714669, + "grad_norm": 1.643127679824829, + "learning_rate": 0.00016523678792038435, + "loss": 0.2858, + "step": 33430 + }, + { + "epoch": 0.5216197666437886, + "grad_norm": 7.096631050109863, + "learning_rate": 0.0001652263888022296, + "loss": 0.4735, + "step": 33440 + }, + { + "epoch": 0.5217757534161103, + "grad_norm": 0.7280219197273254, + "learning_rate": 0.0001652159896840748, + "loss": 0.1841, + "step": 33450 + }, + { + "epoch": 0.521931740188432, + "grad_norm": 0.9415522813796997, + "learning_rate": 0.00016520559056592003, + "loss": 0.2169, + "step": 33460 + }, + { + "epoch": 0.5220877269607537, + "grad_norm": 2.264862060546875, + "learning_rate": 0.00016519519144776523, + "loss": 0.1902, + "step": 33470 + }, + { + "epoch": 0.5222437137330754, + "grad_norm": 2.4566822052001953, + "learning_rate": 0.00016518479232961047, + "loss": 0.3329, + "step": 33480 + }, + { + "epoch": 0.5223997005053972, + "grad_norm": 0.7330378890037537, + "learning_rate": 0.00016517439321145567, + "loss": 0.2341, + "step": 33490 + }, + { + "epoch": 0.5225556872777188, + "grad_norm": 1.6409542560577393, + "learning_rate": 0.0001651639940933009, + "loss": 0.3077, + "step": 33500 + }, + { + "epoch": 0.5227116740500406, + "grad_norm": 1.0647432804107666, + "learning_rate": 0.0001651535949751461, + "loss": 0.1935, + "step": 33510 + }, + { + "epoch": 0.5228676608223622, + "grad_norm": 0.507420539855957, + "learning_rate": 0.00016514319585699135, + "loss": 0.1314, + "step": 33520 + }, + { + "epoch": 0.523023647594684, + "grad_norm": 1.0064164400100708, + "learning_rate": 0.00016513279673883654, + "loss": 0.344, + "step": 33530 + }, + { + "epoch": 0.5231796343670057, + "grad_norm": 1.1936209201812744, + "learning_rate": 0.0001651223976206818, + "loss": 0.1948, + "step": 33540 + }, + { + "epoch": 0.5233356211393274, + "grad_norm": 0.04992926865816116, + "learning_rate": 0.00016511199850252698, + "loss": 0.1363, + "step": 33550 + }, + { + "epoch": 0.5234916079116491, + "grad_norm": 1.8256275653839111, + "learning_rate": 0.00016510159938437223, + "loss": 0.443, + "step": 33560 + }, + { + "epoch": 0.5236475946839708, + "grad_norm": 4.914908409118652, + "learning_rate": 0.00016509120026621742, + "loss": 0.4654, + "step": 33570 + }, + { + "epoch": 0.5238035814562925, + "grad_norm": 1.1625486612319946, + "learning_rate": 0.00016508080114806267, + "loss": 0.1974, + "step": 33580 + }, + { + "epoch": 0.5239595682286142, + "grad_norm": 6.6192522048950195, + "learning_rate": 0.00016507040202990786, + "loss": 0.2131, + "step": 33590 + }, + { + "epoch": 0.5241155550009359, + "grad_norm": 0.7369065880775452, + "learning_rate": 0.0001650600029117531, + "loss": 0.2371, + "step": 33600 + }, + { + "epoch": 0.5242715417732576, + "grad_norm": 1.5238152742385864, + "learning_rate": 0.0001650496037935983, + "loss": 0.1557, + "step": 33610 + }, + { + "epoch": 0.5244275285455793, + "grad_norm": 1.0418007373809814, + "learning_rate": 0.00016503920467544355, + "loss": 0.1878, + "step": 33620 + }, + { + "epoch": 0.524583515317901, + "grad_norm": 0.790117084980011, + "learning_rate": 0.00016502880555728874, + "loss": 0.2195, + "step": 33630 + }, + { + "epoch": 0.5247395020902228, + "grad_norm": 1.6712257862091064, + "learning_rate": 0.000165018406439134, + "loss": 0.1602, + "step": 33640 + }, + { + "epoch": 0.5248954888625444, + "grad_norm": 0.19236230850219727, + "learning_rate": 0.00016500800732097918, + "loss": 0.2526, + "step": 33650 + }, + { + "epoch": 0.5250514756348662, + "grad_norm": 1.3519701957702637, + "learning_rate": 0.00016499760820282443, + "loss": 0.4686, + "step": 33660 + }, + { + "epoch": 0.5252074624071879, + "grad_norm": 1.694342851638794, + "learning_rate": 0.00016498720908466962, + "loss": 0.1859, + "step": 33670 + }, + { + "epoch": 0.5253634491795096, + "grad_norm": 5.225239276885986, + "learning_rate": 0.00016497680996651487, + "loss": 0.2944, + "step": 33680 + }, + { + "epoch": 0.5255194359518313, + "grad_norm": 2.0208842754364014, + "learning_rate": 0.00016496641084836006, + "loss": 0.2421, + "step": 33690 + }, + { + "epoch": 0.525675422724153, + "grad_norm": 0.7954996228218079, + "learning_rate": 0.0001649560117302053, + "loss": 0.2853, + "step": 33700 + }, + { + "epoch": 0.5258314094964747, + "grad_norm": 2.296086072921753, + "learning_rate": 0.0001649456126120505, + "loss": 0.1333, + "step": 33710 + }, + { + "epoch": 0.5259873962687964, + "grad_norm": 1.1779128313064575, + "learning_rate": 0.00016493521349389574, + "loss": 0.1066, + "step": 33720 + }, + { + "epoch": 0.5261433830411181, + "grad_norm": 0.1756065934896469, + "learning_rate": 0.00016492481437574094, + "loss": 0.1352, + "step": 33730 + }, + { + "epoch": 0.5262993698134398, + "grad_norm": 0.13100725412368774, + "learning_rate": 0.00016491441525758618, + "loss": 0.2399, + "step": 33740 + }, + { + "epoch": 0.5264553565857615, + "grad_norm": 5.532008171081543, + "learning_rate": 0.00016490401613943138, + "loss": 0.2896, + "step": 33750 + }, + { + "epoch": 0.5266113433580832, + "grad_norm": 1.319886565208435, + "learning_rate": 0.00016489361702127662, + "loss": 0.3275, + "step": 33760 + }, + { + "epoch": 0.5267673301304049, + "grad_norm": 1.5550974607467651, + "learning_rate": 0.00016488321790312182, + "loss": 0.2677, + "step": 33770 + }, + { + "epoch": 0.5269233169027266, + "grad_norm": 1.8936737775802612, + "learning_rate": 0.00016487281878496706, + "loss": 0.1955, + "step": 33780 + }, + { + "epoch": 0.5270793036750484, + "grad_norm": 0.3653401732444763, + "learning_rate": 0.00016486241966681226, + "loss": 0.0723, + "step": 33790 + }, + { + "epoch": 0.52723529044737, + "grad_norm": 2.861341714859009, + "learning_rate": 0.00016485202054865747, + "loss": 0.2412, + "step": 33800 + }, + { + "epoch": 0.5273912772196918, + "grad_norm": 1.5291428565979004, + "learning_rate": 0.0001648416214305027, + "loss": 0.0871, + "step": 33810 + }, + { + "epoch": 0.5275472639920135, + "grad_norm": 1.0372581481933594, + "learning_rate": 0.00016483122231234791, + "loss": 0.4705, + "step": 33820 + }, + { + "epoch": 0.5277032507643352, + "grad_norm": 1.1943141222000122, + "learning_rate": 0.00016482082319419313, + "loss": 0.2848, + "step": 33830 + }, + { + "epoch": 0.5278592375366569, + "grad_norm": 1.9008225202560425, + "learning_rate": 0.00016481042407603835, + "loss": 0.139, + "step": 33840 + }, + { + "epoch": 0.5280152243089786, + "grad_norm": 2.132089138031006, + "learning_rate": 0.00016480002495788357, + "loss": 0.2119, + "step": 33850 + }, + { + "epoch": 0.5281712110813003, + "grad_norm": 0.24524426460266113, + "learning_rate": 0.0001647896258397288, + "loss": 0.1643, + "step": 33860 + }, + { + "epoch": 0.528327197853622, + "grad_norm": 1.6469637155532837, + "learning_rate": 0.000164779226721574, + "loss": 0.363, + "step": 33870 + }, + { + "epoch": 0.5284831846259437, + "grad_norm": 0.8767328858375549, + "learning_rate": 0.00016476882760341923, + "loss": 0.2632, + "step": 33880 + }, + { + "epoch": 0.5286391713982654, + "grad_norm": 0.06347586214542389, + "learning_rate": 0.00016475842848526445, + "loss": 0.3204, + "step": 33890 + }, + { + "epoch": 0.5287951581705871, + "grad_norm": 0.09782540798187256, + "learning_rate": 0.00016474802936710967, + "loss": 0.2211, + "step": 33900 + }, + { + "epoch": 0.5289511449429088, + "grad_norm": 3.2998859882354736, + "learning_rate": 0.0001647376302489549, + "loss": 0.265, + "step": 33910 + }, + { + "epoch": 0.5291071317152305, + "grad_norm": 0.43594226241111755, + "learning_rate": 0.0001647272311308001, + "loss": 0.194, + "step": 33920 + }, + { + "epoch": 0.5292631184875523, + "grad_norm": 1.5166605710983276, + "learning_rate": 0.00016471683201264533, + "loss": 0.2675, + "step": 33930 + }, + { + "epoch": 0.529419105259874, + "grad_norm": 0.6056640148162842, + "learning_rate": 0.00016470643289449055, + "loss": 0.0576, + "step": 33940 + }, + { + "epoch": 0.5295750920321957, + "grad_norm": 0.25410348176956177, + "learning_rate": 0.00016469603377633577, + "loss": 0.1124, + "step": 33950 + }, + { + "epoch": 0.5297310788045174, + "grad_norm": 1.770642876625061, + "learning_rate": 0.000164685634658181, + "loss": 0.3295, + "step": 33960 + }, + { + "epoch": 0.5298870655768391, + "grad_norm": 0.0607205331325531, + "learning_rate": 0.0001646752355400262, + "loss": 0.2369, + "step": 33970 + }, + { + "epoch": 0.5300430523491608, + "grad_norm": 0.5557095408439636, + "learning_rate": 0.00016466483642187143, + "loss": 0.1681, + "step": 33980 + }, + { + "epoch": 0.5301990391214825, + "grad_norm": 0.5192957520484924, + "learning_rate": 0.00016465443730371665, + "loss": 0.2649, + "step": 33990 + }, + { + "epoch": 0.5303550258938042, + "grad_norm": 0.04804835096001625, + "learning_rate": 0.00016464403818556187, + "loss": 0.2639, + "step": 34000 + }, + { + "epoch": 0.5305110126661259, + "grad_norm": 0.02673129364848137, + "learning_rate": 0.0001646336390674071, + "loss": 0.169, + "step": 34010 + }, + { + "epoch": 0.5306669994384476, + "grad_norm": 1.0084244012832642, + "learning_rate": 0.0001646232399492523, + "loss": 0.353, + "step": 34020 + }, + { + "epoch": 0.5308229862107693, + "grad_norm": 2.2202091217041016, + "learning_rate": 0.00016461284083109753, + "loss": 0.3206, + "step": 34030 + }, + { + "epoch": 0.530978972983091, + "grad_norm": 0.5573744773864746, + "learning_rate": 0.00016460244171294275, + "loss": 0.231, + "step": 34040 + }, + { + "epoch": 0.5311349597554127, + "grad_norm": 0.39700084924697876, + "learning_rate": 0.00016459204259478797, + "loss": 0.1412, + "step": 34050 + }, + { + "epoch": 0.5312909465277345, + "grad_norm": 2.582963228225708, + "learning_rate": 0.00016458164347663319, + "loss": 0.3334, + "step": 34060 + }, + { + "epoch": 0.5314469333000561, + "grad_norm": 0.22781169414520264, + "learning_rate": 0.0001645712443584784, + "loss": 0.2767, + "step": 34070 + }, + { + "epoch": 0.5316029200723779, + "grad_norm": 1.2051042318344116, + "learning_rate": 0.00016456084524032362, + "loss": 0.2045, + "step": 34080 + }, + { + "epoch": 0.5317589068446996, + "grad_norm": 0.42760100960731506, + "learning_rate": 0.00016455044612216884, + "loss": 0.1151, + "step": 34090 + }, + { + "epoch": 0.5319148936170213, + "grad_norm": 0.14440476894378662, + "learning_rate": 0.00016454004700401406, + "loss": 0.2386, + "step": 34100 + }, + { + "epoch": 0.532070880389343, + "grad_norm": 2.2777981758117676, + "learning_rate": 0.00016452964788585928, + "loss": 0.2009, + "step": 34110 + }, + { + "epoch": 0.5322268671616647, + "grad_norm": 0.9206979274749756, + "learning_rate": 0.0001645192487677045, + "loss": 0.2745, + "step": 34120 + }, + { + "epoch": 0.5323828539339864, + "grad_norm": 1.6947574615478516, + "learning_rate": 0.00016450884964954972, + "loss": 0.2584, + "step": 34130 + }, + { + "epoch": 0.5325388407063081, + "grad_norm": 0.401444673538208, + "learning_rate": 0.00016449845053139494, + "loss": 0.2218, + "step": 34140 + }, + { + "epoch": 0.5326948274786298, + "grad_norm": 0.08261553198099136, + "learning_rate": 0.00016448805141324016, + "loss": 0.2775, + "step": 34150 + }, + { + "epoch": 0.5328508142509515, + "grad_norm": 0.1017974391579628, + "learning_rate": 0.00016447765229508538, + "loss": 0.2095, + "step": 34160 + }, + { + "epoch": 0.5330068010232732, + "grad_norm": 1.3759571313858032, + "learning_rate": 0.0001644672531769306, + "loss": 0.2643, + "step": 34170 + }, + { + "epoch": 0.5331627877955949, + "grad_norm": 1.2654389142990112, + "learning_rate": 0.00016445685405877582, + "loss": 0.2949, + "step": 34180 + }, + { + "epoch": 0.5333187745679167, + "grad_norm": 1.5481843948364258, + "learning_rate": 0.00016444645494062104, + "loss": 0.1264, + "step": 34190 + }, + { + "epoch": 0.5334747613402383, + "grad_norm": 1.8094528913497925, + "learning_rate": 0.00016443605582246626, + "loss": 0.2727, + "step": 34200 + }, + { + "epoch": 0.5336307481125601, + "grad_norm": 2.224538564682007, + "learning_rate": 0.00016442565670431148, + "loss": 0.3096, + "step": 34210 + }, + { + "epoch": 0.5337867348848817, + "grad_norm": 0.6375226974487305, + "learning_rate": 0.0001644152575861567, + "loss": 0.2251, + "step": 34220 + }, + { + "epoch": 0.5339427216572035, + "grad_norm": 3.727106809616089, + "learning_rate": 0.00016440485846800192, + "loss": 0.4374, + "step": 34230 + }, + { + "epoch": 0.5340987084295252, + "grad_norm": 0.13345426321029663, + "learning_rate": 0.00016439445934984714, + "loss": 0.2011, + "step": 34240 + }, + { + "epoch": 0.5342546952018469, + "grad_norm": 2.1658668518066406, + "learning_rate": 0.00016438406023169236, + "loss": 0.2457, + "step": 34250 + }, + { + "epoch": 0.5344106819741686, + "grad_norm": 19.238407135009766, + "learning_rate": 0.00016437366111353758, + "loss": 0.2756, + "step": 34260 + }, + { + "epoch": 0.5345666687464903, + "grad_norm": 1.0292778015136719, + "learning_rate": 0.0001643632619953828, + "loss": 0.1646, + "step": 34270 + }, + { + "epoch": 0.534722655518812, + "grad_norm": 0.9372987747192383, + "learning_rate": 0.00016435286287722802, + "loss": 0.2762, + "step": 34280 + }, + { + "epoch": 0.5348786422911337, + "grad_norm": 0.3918002247810364, + "learning_rate": 0.00016434246375907324, + "loss": 0.1406, + "step": 34290 + }, + { + "epoch": 0.5350346290634554, + "grad_norm": 1.3518732786178589, + "learning_rate": 0.00016433206464091846, + "loss": 0.2992, + "step": 34300 + }, + { + "epoch": 0.5351906158357771, + "grad_norm": 0.73117595911026, + "learning_rate": 0.00016432166552276368, + "loss": 0.1172, + "step": 34310 + }, + { + "epoch": 0.5353466026080989, + "grad_norm": 0.06655958294868469, + "learning_rate": 0.0001643112664046089, + "loss": 0.1332, + "step": 34320 + }, + { + "epoch": 0.5355025893804205, + "grad_norm": 3.2587168216705322, + "learning_rate": 0.00016430086728645412, + "loss": 0.2052, + "step": 34330 + }, + { + "epoch": 0.5356585761527423, + "grad_norm": 2.6668431758880615, + "learning_rate": 0.00016429046816829934, + "loss": 0.1295, + "step": 34340 + }, + { + "epoch": 0.5358145629250639, + "grad_norm": 1.316080927848816, + "learning_rate": 0.00016428006905014456, + "loss": 0.2072, + "step": 34350 + }, + { + "epoch": 0.5359705496973857, + "grad_norm": 5.144528388977051, + "learning_rate": 0.00016426966993198977, + "loss": 0.1934, + "step": 34360 + }, + { + "epoch": 0.5361265364697073, + "grad_norm": 1.3529599905014038, + "learning_rate": 0.000164259270813835, + "loss": 0.2077, + "step": 34370 + }, + { + "epoch": 0.5362825232420291, + "grad_norm": 0.9132925868034363, + "learning_rate": 0.00016424887169568021, + "loss": 0.2404, + "step": 34380 + }, + { + "epoch": 0.5364385100143508, + "grad_norm": 1.520033597946167, + "learning_rate": 0.00016423847257752543, + "loss": 0.1765, + "step": 34390 + }, + { + "epoch": 0.5365944967866725, + "grad_norm": 2.343975305557251, + "learning_rate": 0.00016422807345937065, + "loss": 0.1959, + "step": 34400 + }, + { + "epoch": 0.5367504835589942, + "grad_norm": 0.5855118632316589, + "learning_rate": 0.00016421767434121587, + "loss": 0.0848, + "step": 34410 + }, + { + "epoch": 0.5369064703313159, + "grad_norm": 1.9108648300170898, + "learning_rate": 0.0001642072752230611, + "loss": 0.2406, + "step": 34420 + }, + { + "epoch": 0.5370624571036376, + "grad_norm": 1.0323792695999146, + "learning_rate": 0.0001641968761049063, + "loss": 0.2298, + "step": 34430 + }, + { + "epoch": 0.5372184438759593, + "grad_norm": 0.17435620725154877, + "learning_rate": 0.00016418647698675153, + "loss": 0.1396, + "step": 34440 + }, + { + "epoch": 0.537374430648281, + "grad_norm": 0.3526577353477478, + "learning_rate": 0.00016417607786859675, + "loss": 0.172, + "step": 34450 + }, + { + "epoch": 0.5375304174206027, + "grad_norm": 2.73685884475708, + "learning_rate": 0.00016416567875044197, + "loss": 0.1167, + "step": 34460 + }, + { + "epoch": 0.5376864041929245, + "grad_norm": 2.765693426132202, + "learning_rate": 0.0001641552796322872, + "loss": 0.1356, + "step": 34470 + }, + { + "epoch": 0.5378423909652461, + "grad_norm": 0.47628021240234375, + "learning_rate": 0.0001641448805141324, + "loss": 0.3104, + "step": 34480 + }, + { + "epoch": 0.5379983777375679, + "grad_norm": 1.1307590007781982, + "learning_rate": 0.00016413448139597763, + "loss": 0.3247, + "step": 34490 + }, + { + "epoch": 0.5381543645098895, + "grad_norm": 0.07136381417512894, + "learning_rate": 0.00016412408227782285, + "loss": 0.319, + "step": 34500 + }, + { + "epoch": 0.5383103512822113, + "grad_norm": 1.7850221395492554, + "learning_rate": 0.00016411368315966807, + "loss": 0.2515, + "step": 34510 + }, + { + "epoch": 0.5384663380545329, + "grad_norm": 1.3870742321014404, + "learning_rate": 0.0001641032840415133, + "loss": 0.256, + "step": 34520 + }, + { + "epoch": 0.5386223248268547, + "grad_norm": 1.1329221725463867, + "learning_rate": 0.0001640928849233585, + "loss": 0.2255, + "step": 34530 + }, + { + "epoch": 0.5387783115991764, + "grad_norm": 0.23236137628555298, + "learning_rate": 0.00016408248580520373, + "loss": 0.4053, + "step": 34540 + }, + { + "epoch": 0.5389342983714981, + "grad_norm": 3.350924253463745, + "learning_rate": 0.00016407208668704895, + "loss": 0.2816, + "step": 34550 + }, + { + "epoch": 0.5390902851438198, + "grad_norm": 1.7459170818328857, + "learning_rate": 0.00016406168756889417, + "loss": 0.0834, + "step": 34560 + }, + { + "epoch": 0.5392462719161415, + "grad_norm": 0.89351487159729, + "learning_rate": 0.0001640512884507394, + "loss": 0.1146, + "step": 34570 + }, + { + "epoch": 0.5394022586884633, + "grad_norm": 0.49410831928253174, + "learning_rate": 0.0001640408893325846, + "loss": 0.2319, + "step": 34580 + }, + { + "epoch": 0.5395582454607849, + "grad_norm": 0.009971237741410732, + "learning_rate": 0.00016403049021442983, + "loss": 0.1674, + "step": 34590 + }, + { + "epoch": 0.5397142322331067, + "grad_norm": 1.420511245727539, + "learning_rate": 0.00016402009109627505, + "loss": 0.1469, + "step": 34600 + }, + { + "epoch": 0.5398702190054283, + "grad_norm": 2.5686769485473633, + "learning_rate": 0.00016400969197812027, + "loss": 0.2607, + "step": 34610 + }, + { + "epoch": 0.5400262057777501, + "grad_norm": 2.5808913707733154, + "learning_rate": 0.00016399929285996549, + "loss": 0.4412, + "step": 34620 + }, + { + "epoch": 0.5401821925500717, + "grad_norm": 0.2778591811656952, + "learning_rate": 0.0001639888937418107, + "loss": 0.1563, + "step": 34630 + }, + { + "epoch": 0.5403381793223935, + "grad_norm": 1.6448099613189697, + "learning_rate": 0.00016397849462365592, + "loss": 0.2484, + "step": 34640 + }, + { + "epoch": 0.5404941660947151, + "grad_norm": 0.003213417250663042, + "learning_rate": 0.00016396809550550114, + "loss": 0.1582, + "step": 34650 + }, + { + "epoch": 0.5406501528670369, + "grad_norm": 2.3433990478515625, + "learning_rate": 0.00016395769638734636, + "loss": 0.2774, + "step": 34660 + }, + { + "epoch": 0.5408061396393585, + "grad_norm": 0.09544426947832108, + "learning_rate": 0.00016394729726919158, + "loss": 0.1242, + "step": 34670 + }, + { + "epoch": 0.5409621264116803, + "grad_norm": 1.8934235572814941, + "learning_rate": 0.0001639368981510368, + "loss": 0.1636, + "step": 34680 + }, + { + "epoch": 0.541118113184002, + "grad_norm": 2.8528554439544678, + "learning_rate": 0.00016392649903288202, + "loss": 0.2524, + "step": 34690 + }, + { + "epoch": 0.5412740999563237, + "grad_norm": 2.837836980819702, + "learning_rate": 0.00016391609991472722, + "loss": 0.1808, + "step": 34700 + }, + { + "epoch": 0.5414300867286455, + "grad_norm": 4.349740028381348, + "learning_rate": 0.00016390570079657246, + "loss": 0.12, + "step": 34710 + }, + { + "epoch": 0.5415860735009671, + "grad_norm": 0.6011945605278015, + "learning_rate": 0.00016389530167841765, + "loss": 0.2865, + "step": 34720 + }, + { + "epoch": 0.5417420602732889, + "grad_norm": 1.2079488039016724, + "learning_rate": 0.0001638849025602629, + "loss": 0.125, + "step": 34730 + }, + { + "epoch": 0.5418980470456105, + "grad_norm": 1.9549164772033691, + "learning_rate": 0.0001638745034421081, + "loss": 0.3285, + "step": 34740 + }, + { + "epoch": 0.5420540338179323, + "grad_norm": 0.6969407200813293, + "learning_rate": 0.00016386410432395334, + "loss": 0.3267, + "step": 34750 + }, + { + "epoch": 0.5422100205902539, + "grad_norm": 1.0561965703964233, + "learning_rate": 0.00016385370520579853, + "loss": 0.1463, + "step": 34760 + }, + { + "epoch": 0.5423660073625757, + "grad_norm": 1.8208931684494019, + "learning_rate": 0.00016384330608764378, + "loss": 0.1713, + "step": 34770 + }, + { + "epoch": 0.5425219941348973, + "grad_norm": 0.6349910497665405, + "learning_rate": 0.000163832906969489, + "loss": 0.2244, + "step": 34780 + }, + { + "epoch": 0.5426779809072191, + "grad_norm": 1.1976515054702759, + "learning_rate": 0.00016382250785133422, + "loss": 0.3407, + "step": 34790 + }, + { + "epoch": 0.5428339676795407, + "grad_norm": 0.5970319509506226, + "learning_rate": 0.00016381210873317944, + "loss": 0.1419, + "step": 34800 + }, + { + "epoch": 0.5429899544518625, + "grad_norm": 0.14623159170150757, + "learning_rate": 0.00016380170961502466, + "loss": 0.14, + "step": 34810 + }, + { + "epoch": 0.5431459412241841, + "grad_norm": 1.4332351684570312, + "learning_rate": 0.00016379131049686988, + "loss": 0.1093, + "step": 34820 + }, + { + "epoch": 0.5433019279965059, + "grad_norm": 1.1443063020706177, + "learning_rate": 0.0001637809113787151, + "loss": 0.3676, + "step": 34830 + }, + { + "epoch": 0.5434579147688277, + "grad_norm": 1.3897461891174316, + "learning_rate": 0.00016377051226056032, + "loss": 0.3713, + "step": 34840 + }, + { + "epoch": 0.5436139015411493, + "grad_norm": 1.7773199081420898, + "learning_rate": 0.00016376011314240554, + "loss": 0.2249, + "step": 34850 + }, + { + "epoch": 0.5437698883134711, + "grad_norm": 0.8524389266967773, + "learning_rate": 0.00016374971402425076, + "loss": 0.2158, + "step": 34860 + }, + { + "epoch": 0.5439258750857927, + "grad_norm": 0.5572389960289001, + "learning_rate": 0.00016373931490609598, + "loss": 0.215, + "step": 34870 + }, + { + "epoch": 0.5440818618581145, + "grad_norm": 3.0238709449768066, + "learning_rate": 0.0001637289157879412, + "loss": 0.3765, + "step": 34880 + }, + { + "epoch": 0.5442378486304361, + "grad_norm": 1.1655421257019043, + "learning_rate": 0.00016371851666978642, + "loss": 0.2491, + "step": 34890 + }, + { + "epoch": 0.5443938354027579, + "grad_norm": 0.49492090940475464, + "learning_rate": 0.00016370811755163164, + "loss": 0.1287, + "step": 34900 + }, + { + "epoch": 0.5445498221750795, + "grad_norm": 0.2732921540737152, + "learning_rate": 0.00016369771843347685, + "loss": 0.2154, + "step": 34910 + }, + { + "epoch": 0.5447058089474013, + "grad_norm": 2.5807769298553467, + "learning_rate": 0.00016368731931532207, + "loss": 0.2032, + "step": 34920 + }, + { + "epoch": 0.5448617957197229, + "grad_norm": 0.5044315457344055, + "learning_rate": 0.0001636769201971673, + "loss": 0.2595, + "step": 34930 + }, + { + "epoch": 0.5450177824920447, + "grad_norm": 2.1248972415924072, + "learning_rate": 0.00016366652107901251, + "loss": 0.5651, + "step": 34940 + }, + { + "epoch": 0.5451737692643663, + "grad_norm": 0.8391468524932861, + "learning_rate": 0.00016365612196085773, + "loss": 0.3028, + "step": 34950 + }, + { + "epoch": 0.5453297560366881, + "grad_norm": 0.36081477999687195, + "learning_rate": 0.00016364572284270295, + "loss": 0.1073, + "step": 34960 + }, + { + "epoch": 0.5454857428090097, + "grad_norm": 3.175804376602173, + "learning_rate": 0.00016363532372454817, + "loss": 0.1889, + "step": 34970 + }, + { + "epoch": 0.5456417295813315, + "grad_norm": 0.20647937059402466, + "learning_rate": 0.0001636249246063934, + "loss": 0.2416, + "step": 34980 + }, + { + "epoch": 0.5457977163536533, + "grad_norm": 2.3525943756103516, + "learning_rate": 0.0001636145254882386, + "loss": 0.3179, + "step": 34990 + }, + { + "epoch": 0.5459537031259749, + "grad_norm": 0.2641216814517975, + "learning_rate": 0.00016360412637008383, + "loss": 0.2157, + "step": 35000 + }, + { + "epoch": 0.5461096898982967, + "grad_norm": 0.09906073659658432, + "learning_rate": 0.00016359372725192905, + "loss": 0.2576, + "step": 35010 + }, + { + "epoch": 0.5462656766706183, + "grad_norm": 0.13971920311450958, + "learning_rate": 0.00016358332813377427, + "loss": 0.1298, + "step": 35020 + }, + { + "epoch": 0.5464216634429401, + "grad_norm": 1.4535638093948364, + "learning_rate": 0.0001635729290156195, + "loss": 0.2357, + "step": 35030 + }, + { + "epoch": 0.5465776502152617, + "grad_norm": 0.9107828736305237, + "learning_rate": 0.0001635625298974647, + "loss": 0.3049, + "step": 35040 + }, + { + "epoch": 0.5467336369875835, + "grad_norm": 0.0377386212348938, + "learning_rate": 0.00016355213077930993, + "loss": 0.2066, + "step": 35050 + }, + { + "epoch": 0.5468896237599051, + "grad_norm": 2.7800869941711426, + "learning_rate": 0.00016354173166115515, + "loss": 0.2974, + "step": 35060 + }, + { + "epoch": 0.5470456105322269, + "grad_norm": 1.6247998476028442, + "learning_rate": 0.00016353133254300037, + "loss": 0.3822, + "step": 35070 + }, + { + "epoch": 0.5472015973045485, + "grad_norm": 1.6479015350341797, + "learning_rate": 0.0001635209334248456, + "loss": 0.3401, + "step": 35080 + }, + { + "epoch": 0.5473575840768703, + "grad_norm": 0.1879737675189972, + "learning_rate": 0.0001635105343066908, + "loss": 0.1549, + "step": 35090 + }, + { + "epoch": 0.5475135708491919, + "grad_norm": 0.39355361461639404, + "learning_rate": 0.00016350013518853603, + "loss": 0.2032, + "step": 35100 + }, + { + "epoch": 0.5476695576215137, + "grad_norm": 2.9976983070373535, + "learning_rate": 0.00016348973607038125, + "loss": 0.3867, + "step": 35110 + }, + { + "epoch": 0.5478255443938354, + "grad_norm": 0.8242707252502441, + "learning_rate": 0.00016347933695222647, + "loss": 0.3601, + "step": 35120 + }, + { + "epoch": 0.5479815311661571, + "grad_norm": 0.874955415725708, + "learning_rate": 0.0001634689378340717, + "loss": 0.2709, + "step": 35130 + }, + { + "epoch": 0.5481375179384789, + "grad_norm": 0.05980971083045006, + "learning_rate": 0.0001634585387159169, + "loss": 0.2646, + "step": 35140 + }, + { + "epoch": 0.5482935047108005, + "grad_norm": 1.8437328338623047, + "learning_rate": 0.0001634481395977621, + "loss": 0.1965, + "step": 35150 + }, + { + "epoch": 0.5484494914831223, + "grad_norm": 0.4142405688762665, + "learning_rate": 0.00016343774047960735, + "loss": 0.1756, + "step": 35160 + }, + { + "epoch": 0.5486054782554439, + "grad_norm": 4.836324214935303, + "learning_rate": 0.00016342734136145254, + "loss": 0.2534, + "step": 35170 + }, + { + "epoch": 0.5487614650277657, + "grad_norm": 1.4933065176010132, + "learning_rate": 0.00016341694224329779, + "loss": 0.2355, + "step": 35180 + }, + { + "epoch": 0.5489174518000873, + "grad_norm": 1.9468894004821777, + "learning_rate": 0.00016340654312514298, + "loss": 0.3113, + "step": 35190 + }, + { + "epoch": 0.5490734385724091, + "grad_norm": 1.012710452079773, + "learning_rate": 0.00016339614400698822, + "loss": 0.2635, + "step": 35200 + }, + { + "epoch": 0.5492294253447307, + "grad_norm": 1.5551140308380127, + "learning_rate": 0.00016338574488883342, + "loss": 0.2951, + "step": 35210 + }, + { + "epoch": 0.5493854121170525, + "grad_norm": 1.8613696098327637, + "learning_rate": 0.00016337534577067866, + "loss": 0.2832, + "step": 35220 + }, + { + "epoch": 0.5495413988893741, + "grad_norm": 1.6748839616775513, + "learning_rate": 0.00016336494665252386, + "loss": 0.2266, + "step": 35230 + }, + { + "epoch": 0.5496973856616959, + "grad_norm": 0.4094032347202301, + "learning_rate": 0.0001633545475343691, + "loss": 0.3497, + "step": 35240 + }, + { + "epoch": 0.5498533724340176, + "grad_norm": 1.6414631605148315, + "learning_rate": 0.0001633441484162143, + "loss": 0.1763, + "step": 35250 + }, + { + "epoch": 0.5500093592063393, + "grad_norm": 1.6380645036697388, + "learning_rate": 0.00016333374929805954, + "loss": 0.1355, + "step": 35260 + }, + { + "epoch": 0.550165345978661, + "grad_norm": 0.7325630187988281, + "learning_rate": 0.00016332335017990473, + "loss": 0.3105, + "step": 35270 + }, + { + "epoch": 0.5503213327509827, + "grad_norm": 1.6015644073486328, + "learning_rate": 0.00016331295106174998, + "loss": 0.2311, + "step": 35280 + }, + { + "epoch": 0.5504773195233045, + "grad_norm": 0.9224210381507874, + "learning_rate": 0.00016330255194359517, + "loss": 0.2746, + "step": 35290 + }, + { + "epoch": 0.5506333062956261, + "grad_norm": 1.30025315284729, + "learning_rate": 0.00016329215282544042, + "loss": 0.2066, + "step": 35300 + }, + { + "epoch": 0.5507892930679479, + "grad_norm": 0.481125146150589, + "learning_rate": 0.0001632817537072856, + "loss": 0.1778, + "step": 35310 + }, + { + "epoch": 0.5509452798402695, + "grad_norm": 0.2709486782550812, + "learning_rate": 0.00016327135458913086, + "loss": 0.1131, + "step": 35320 + }, + { + "epoch": 0.5511012666125913, + "grad_norm": 0.6248563528060913, + "learning_rate": 0.00016326095547097605, + "loss": 0.4356, + "step": 35330 + }, + { + "epoch": 0.5512572533849129, + "grad_norm": 3.0947647094726562, + "learning_rate": 0.0001632505563528213, + "loss": 0.3072, + "step": 35340 + }, + { + "epoch": 0.5514132401572347, + "grad_norm": 1.009535312652588, + "learning_rate": 0.0001632401572346665, + "loss": 0.1808, + "step": 35350 + }, + { + "epoch": 0.5515692269295563, + "grad_norm": 0.042604975402355194, + "learning_rate": 0.00016322975811651174, + "loss": 0.1198, + "step": 35360 + }, + { + "epoch": 0.5517252137018781, + "grad_norm": 1.8652396202087402, + "learning_rate": 0.00016321935899835693, + "loss": 0.2176, + "step": 35370 + }, + { + "epoch": 0.5518812004741998, + "grad_norm": 4.6097187995910645, + "learning_rate": 0.00016320895988020218, + "loss": 0.2132, + "step": 35380 + }, + { + "epoch": 0.5520371872465215, + "grad_norm": 1.3615522384643555, + "learning_rate": 0.00016319856076204737, + "loss": 0.2751, + "step": 35390 + }, + { + "epoch": 0.5521931740188432, + "grad_norm": 0.8844773769378662, + "learning_rate": 0.00016318816164389262, + "loss": 0.22, + "step": 35400 + }, + { + "epoch": 0.5523491607911649, + "grad_norm": 0.5413331985473633, + "learning_rate": 0.0001631777625257378, + "loss": 0.2159, + "step": 35410 + }, + { + "epoch": 0.5525051475634866, + "grad_norm": 1.5456678867340088, + "learning_rate": 0.00016316736340758306, + "loss": 0.2244, + "step": 35420 + }, + { + "epoch": 0.5526611343358083, + "grad_norm": 2.1405861377716064, + "learning_rate": 0.00016315696428942825, + "loss": 0.446, + "step": 35430 + }, + { + "epoch": 0.5528171211081301, + "grad_norm": 1.4269858598709106, + "learning_rate": 0.0001631465651712735, + "loss": 0.3599, + "step": 35440 + }, + { + "epoch": 0.5529731078804517, + "grad_norm": 1.8682516813278198, + "learning_rate": 0.0001631361660531187, + "loss": 0.115, + "step": 35450 + }, + { + "epoch": 0.5531290946527735, + "grad_norm": 1.0175774097442627, + "learning_rate": 0.00016312576693496394, + "loss": 0.4003, + "step": 35460 + }, + { + "epoch": 0.5532850814250951, + "grad_norm": 2.4369170665740967, + "learning_rate": 0.00016311536781680913, + "loss": 0.2489, + "step": 35470 + }, + { + "epoch": 0.5534410681974169, + "grad_norm": 1.886022686958313, + "learning_rate": 0.00016310496869865437, + "loss": 0.4201, + "step": 35480 + }, + { + "epoch": 0.5535970549697385, + "grad_norm": 1.5811959505081177, + "learning_rate": 0.00016309456958049957, + "loss": 0.4286, + "step": 35490 + }, + { + "epoch": 0.5537530417420603, + "grad_norm": 1.971110224723816, + "learning_rate": 0.00016308417046234481, + "loss": 0.3599, + "step": 35500 + }, + { + "epoch": 0.553909028514382, + "grad_norm": 0.0335380844771862, + "learning_rate": 0.00016307377134419, + "loss": 0.3357, + "step": 35510 + }, + { + "epoch": 0.5540650152867037, + "grad_norm": 1.8519577980041504, + "learning_rate": 0.00016306337222603525, + "loss": 0.3126, + "step": 35520 + }, + { + "epoch": 0.5542210020590254, + "grad_norm": 2.1563379764556885, + "learning_rate": 0.00016305297310788045, + "loss": 0.1123, + "step": 35530 + }, + { + "epoch": 0.5543769888313471, + "grad_norm": 1.64332914352417, + "learning_rate": 0.0001630425739897257, + "loss": 0.2853, + "step": 35540 + }, + { + "epoch": 0.5545329756036688, + "grad_norm": 0.061150554567575455, + "learning_rate": 0.00016303217487157088, + "loss": 0.1943, + "step": 35550 + }, + { + "epoch": 0.5546889623759905, + "grad_norm": 1.2701060771942139, + "learning_rate": 0.00016302177575341613, + "loss": 0.2304, + "step": 35560 + }, + { + "epoch": 0.5548449491483122, + "grad_norm": 2.424860715866089, + "learning_rate": 0.00016301137663526132, + "loss": 0.2128, + "step": 35570 + }, + { + "epoch": 0.5550009359206339, + "grad_norm": 0.6803575158119202, + "learning_rate": 0.00016300097751710657, + "loss": 0.1145, + "step": 35580 + }, + { + "epoch": 0.5551569226929557, + "grad_norm": 1.2855092287063599, + "learning_rate": 0.00016299057839895176, + "loss": 0.2448, + "step": 35590 + }, + { + "epoch": 0.5553129094652773, + "grad_norm": 2.6340911388397217, + "learning_rate": 0.00016298017928079698, + "loss": 0.1856, + "step": 35600 + }, + { + "epoch": 0.5554688962375991, + "grad_norm": 1.776382327079773, + "learning_rate": 0.0001629697801626422, + "loss": 0.2422, + "step": 35610 + }, + { + "epoch": 0.5556248830099207, + "grad_norm": 3.0746347904205322, + "learning_rate": 0.00016295938104448742, + "loss": 0.1578, + "step": 35620 + }, + { + "epoch": 0.5557808697822425, + "grad_norm": 1.4887659549713135, + "learning_rate": 0.00016294898192633267, + "loss": 0.2385, + "step": 35630 + }, + { + "epoch": 0.5559368565545642, + "grad_norm": 0.5371220707893372, + "learning_rate": 0.00016293858280817786, + "loss": 0.2361, + "step": 35640 + }, + { + "epoch": 0.5560928433268859, + "grad_norm": 0.18604904413223267, + "learning_rate": 0.0001629281836900231, + "loss": 0.0828, + "step": 35650 + }, + { + "epoch": 0.5562488300992076, + "grad_norm": 2.267854690551758, + "learning_rate": 0.0001629177845718683, + "loss": 0.413, + "step": 35660 + }, + { + "epoch": 0.5564048168715293, + "grad_norm": 1.9755452871322632, + "learning_rate": 0.00016290738545371355, + "loss": 0.146, + "step": 35670 + }, + { + "epoch": 0.556560803643851, + "grad_norm": 2.430293321609497, + "learning_rate": 0.00016289698633555874, + "loss": 0.2677, + "step": 35680 + }, + { + "epoch": 0.5567167904161727, + "grad_norm": 0.6319543719291687, + "learning_rate": 0.000162886587217404, + "loss": 0.1481, + "step": 35690 + }, + { + "epoch": 0.5568727771884944, + "grad_norm": 0.17439277470111847, + "learning_rate": 0.00016287618809924918, + "loss": 0.1641, + "step": 35700 + }, + { + "epoch": 0.5570287639608161, + "grad_norm": 0.350175142288208, + "learning_rate": 0.00016286578898109443, + "loss": 0.393, + "step": 35710 + }, + { + "epoch": 0.5571847507331378, + "grad_norm": 2.641941547393799, + "learning_rate": 0.00016285538986293962, + "loss": 0.1907, + "step": 35720 + }, + { + "epoch": 0.5573407375054595, + "grad_norm": 0.1910303384065628, + "learning_rate": 0.00016284499074478487, + "loss": 0.2248, + "step": 35730 + }, + { + "epoch": 0.5574967242777813, + "grad_norm": 1.0180896520614624, + "learning_rate": 0.00016283459162663006, + "loss": 0.1924, + "step": 35740 + }, + { + "epoch": 0.557652711050103, + "grad_norm": 1.2047260999679565, + "learning_rate": 0.0001628241925084753, + "loss": 0.2031, + "step": 35750 + }, + { + "epoch": 0.5578086978224247, + "grad_norm": 1.8182405233383179, + "learning_rate": 0.0001628137933903205, + "loss": 0.1573, + "step": 35760 + }, + { + "epoch": 0.5579646845947464, + "grad_norm": 0.5485963821411133, + "learning_rate": 0.00016280339427216574, + "loss": 0.152, + "step": 35770 + }, + { + "epoch": 0.5581206713670681, + "grad_norm": 0.1674145758152008, + "learning_rate": 0.00016279299515401094, + "loss": 0.2388, + "step": 35780 + }, + { + "epoch": 0.5582766581393898, + "grad_norm": 0.36187657713890076, + "learning_rate": 0.00016278259603585618, + "loss": 0.1618, + "step": 35790 + }, + { + "epoch": 0.5584326449117115, + "grad_norm": 0.013768521137535572, + "learning_rate": 0.00016277219691770138, + "loss": 0.3666, + "step": 35800 + }, + { + "epoch": 0.5585886316840332, + "grad_norm": 2.8602802753448486, + "learning_rate": 0.00016276179779954662, + "loss": 0.2065, + "step": 35810 + }, + { + "epoch": 0.5587446184563549, + "grad_norm": 1.7925455570220947, + "learning_rate": 0.00016275139868139182, + "loss": 0.1729, + "step": 35820 + }, + { + "epoch": 0.5589006052286766, + "grad_norm": 2.910456418991089, + "learning_rate": 0.00016274099956323706, + "loss": 0.3734, + "step": 35830 + }, + { + "epoch": 0.5590565920009983, + "grad_norm": 0.8378308415412903, + "learning_rate": 0.00016273060044508225, + "loss": 0.2255, + "step": 35840 + }, + { + "epoch": 0.55921257877332, + "grad_norm": 0.409534752368927, + "learning_rate": 0.0001627202013269275, + "loss": 0.2046, + "step": 35850 + }, + { + "epoch": 0.5593685655456417, + "grad_norm": 0.49498099088668823, + "learning_rate": 0.0001627098022087727, + "loss": 0.1058, + "step": 35860 + }, + { + "epoch": 0.5595245523179634, + "grad_norm": 1.1761783361434937, + "learning_rate": 0.00016269940309061794, + "loss": 0.1566, + "step": 35870 + }, + { + "epoch": 0.5596805390902851, + "grad_norm": 3.8204751014709473, + "learning_rate": 0.00016268900397246313, + "loss": 0.2646, + "step": 35880 + }, + { + "epoch": 0.5598365258626069, + "grad_norm": 0.9882522225379944, + "learning_rate": 0.00016267860485430838, + "loss": 0.1756, + "step": 35890 + }, + { + "epoch": 0.5599925126349286, + "grad_norm": 1.1832259893417358, + "learning_rate": 0.00016266820573615357, + "loss": 0.1385, + "step": 35900 + }, + { + "epoch": 0.5601484994072503, + "grad_norm": 0.7638296484947205, + "learning_rate": 0.00016265780661799882, + "loss": 0.5679, + "step": 35910 + }, + { + "epoch": 0.560304486179572, + "grad_norm": 0.8551504611968994, + "learning_rate": 0.000162647407499844, + "loss": 0.2981, + "step": 35920 + }, + { + "epoch": 0.5604604729518937, + "grad_norm": 5.499948501586914, + "learning_rate": 0.00016263700838168926, + "loss": 0.1429, + "step": 35930 + }, + { + "epoch": 0.5606164597242154, + "grad_norm": 3.0494496822357178, + "learning_rate": 0.00016262660926353445, + "loss": 0.1322, + "step": 35940 + }, + { + "epoch": 0.5607724464965371, + "grad_norm": 0.2482384592294693, + "learning_rate": 0.0001626162101453797, + "loss": 0.2473, + "step": 35950 + }, + { + "epoch": 0.5609284332688588, + "grad_norm": 0.009750776924192905, + "learning_rate": 0.0001626058110272249, + "loss": 0.0429, + "step": 35960 + }, + { + "epoch": 0.5610844200411805, + "grad_norm": 0.5742604732513428, + "learning_rate": 0.00016259541190907014, + "loss": 0.4565, + "step": 35970 + }, + { + "epoch": 0.5612404068135022, + "grad_norm": 0.7810243368148804, + "learning_rate": 0.00016258501279091533, + "loss": 0.2107, + "step": 35980 + }, + { + "epoch": 0.5613963935858239, + "grad_norm": 2.785747528076172, + "learning_rate": 0.00016257461367276058, + "loss": 0.5087, + "step": 35990 + }, + { + "epoch": 0.5615523803581456, + "grad_norm": 0.8806902766227722, + "learning_rate": 0.00016256421455460577, + "loss": 0.2937, + "step": 36000 + }, + { + "epoch": 0.5617083671304673, + "grad_norm": 1.857373833656311, + "learning_rate": 0.00016255381543645102, + "loss": 0.5042, + "step": 36010 + }, + { + "epoch": 0.561864353902789, + "grad_norm": 0.5333901047706604, + "learning_rate": 0.0001625434163182962, + "loss": 0.1678, + "step": 36020 + }, + { + "epoch": 0.5620203406751108, + "grad_norm": 11.023160934448242, + "learning_rate": 0.00016253301720014145, + "loss": 0.2721, + "step": 36030 + }, + { + "epoch": 0.5621763274474325, + "grad_norm": 2.6214029788970947, + "learning_rate": 0.00016252261808198665, + "loss": 0.1675, + "step": 36040 + }, + { + "epoch": 0.5623323142197542, + "grad_norm": 0.25980064272880554, + "learning_rate": 0.0001625122189638319, + "loss": 0.1832, + "step": 36050 + }, + { + "epoch": 0.5624883009920759, + "grad_norm": 1.3559473752975464, + "learning_rate": 0.0001625018198456771, + "loss": 0.121, + "step": 36060 + }, + { + "epoch": 0.5626442877643976, + "grad_norm": 0.3558153212070465, + "learning_rate": 0.0001624914207275223, + "loss": 0.2797, + "step": 36070 + }, + { + "epoch": 0.5628002745367193, + "grad_norm": 0.6314427256584167, + "learning_rate": 0.00016248102160936753, + "loss": 0.3455, + "step": 36080 + }, + { + "epoch": 0.562956261309041, + "grad_norm": 0.16120101511478424, + "learning_rate": 0.00016247062249121275, + "loss": 0.3669, + "step": 36090 + }, + { + "epoch": 0.5631122480813627, + "grad_norm": 1.041572093963623, + "learning_rate": 0.00016246022337305797, + "loss": 0.2744, + "step": 36100 + }, + { + "epoch": 0.5632682348536844, + "grad_norm": 0.2645890414714813, + "learning_rate": 0.00016244982425490318, + "loss": 0.1487, + "step": 36110 + }, + { + "epoch": 0.5634242216260061, + "grad_norm": 5.698698043823242, + "learning_rate": 0.0001624394251367484, + "loss": 0.173, + "step": 36120 + }, + { + "epoch": 0.5635802083983278, + "grad_norm": 0.3299804627895355, + "learning_rate": 0.00016242902601859362, + "loss": 0.1965, + "step": 36130 + }, + { + "epoch": 0.5637361951706495, + "grad_norm": 1.2476481199264526, + "learning_rate": 0.00016241862690043884, + "loss": 0.4097, + "step": 36140 + }, + { + "epoch": 0.5638921819429712, + "grad_norm": 0.30926263332366943, + "learning_rate": 0.00016240822778228406, + "loss": 0.1416, + "step": 36150 + }, + { + "epoch": 0.564048168715293, + "grad_norm": 0.05872740224003792, + "learning_rate": 0.00016239782866412928, + "loss": 0.1449, + "step": 36160 + }, + { + "epoch": 0.5642041554876146, + "grad_norm": 0.2292211800813675, + "learning_rate": 0.0001623874295459745, + "loss": 0.2557, + "step": 36170 + }, + { + "epoch": 0.5643601422599364, + "grad_norm": 1.7822531461715698, + "learning_rate": 0.00016237703042781972, + "loss": 0.2953, + "step": 36180 + }, + { + "epoch": 0.5645161290322581, + "grad_norm": 2.4908461570739746, + "learning_rate": 0.00016236663130966494, + "loss": 0.3888, + "step": 36190 + }, + { + "epoch": 0.5646721158045798, + "grad_norm": 0.08363594114780426, + "learning_rate": 0.00016235623219151016, + "loss": 0.2293, + "step": 36200 + }, + { + "epoch": 0.5648281025769015, + "grad_norm": 1.4830002784729004, + "learning_rate": 0.00016234583307335538, + "loss": 0.2348, + "step": 36210 + }, + { + "epoch": 0.5649840893492232, + "grad_norm": 1.8443071842193604, + "learning_rate": 0.0001623354339552006, + "loss": 0.3392, + "step": 36220 + }, + { + "epoch": 0.5651400761215449, + "grad_norm": 2.8050875663757324, + "learning_rate": 0.00016232503483704582, + "loss": 0.2426, + "step": 36230 + }, + { + "epoch": 0.5652960628938666, + "grad_norm": 3.6627259254455566, + "learning_rate": 0.00016231463571889104, + "loss": 0.2118, + "step": 36240 + }, + { + "epoch": 0.5654520496661883, + "grad_norm": 0.3443094789981842, + "learning_rate": 0.00016230423660073626, + "loss": 0.2418, + "step": 36250 + }, + { + "epoch": 0.56560803643851, + "grad_norm": 1.7411112785339355, + "learning_rate": 0.00016229383748258148, + "loss": 0.1628, + "step": 36260 + }, + { + "epoch": 0.5657640232108317, + "grad_norm": 1.1321816444396973, + "learning_rate": 0.0001622834383644267, + "loss": 0.2732, + "step": 36270 + }, + { + "epoch": 0.5659200099831534, + "grad_norm": 2.9686615467071533, + "learning_rate": 0.00016227303924627192, + "loss": 0.3036, + "step": 36280 + }, + { + "epoch": 0.5660759967554752, + "grad_norm": 5.03530740737915, + "learning_rate": 0.00016226264012811714, + "loss": 0.2143, + "step": 36290 + }, + { + "epoch": 0.5662319835277968, + "grad_norm": 2.3943281173706055, + "learning_rate": 0.00016225224100996236, + "loss": 0.2629, + "step": 36300 + }, + { + "epoch": 0.5663879703001186, + "grad_norm": 1.812828779220581, + "learning_rate": 0.00016224184189180758, + "loss": 0.3892, + "step": 36310 + }, + { + "epoch": 0.5665439570724402, + "grad_norm": 1.4708483219146729, + "learning_rate": 0.0001622314427736528, + "loss": 0.2118, + "step": 36320 + }, + { + "epoch": 0.566699943844762, + "grad_norm": 8.99713134765625, + "learning_rate": 0.00016222104365549802, + "loss": 0.3185, + "step": 36330 + }, + { + "epoch": 0.5668559306170837, + "grad_norm": 1.7472341060638428, + "learning_rate": 0.00016221064453734324, + "loss": 0.1659, + "step": 36340 + }, + { + "epoch": 0.5670119173894054, + "grad_norm": 0.8059778213500977, + "learning_rate": 0.00016220024541918846, + "loss": 0.0621, + "step": 36350 + }, + { + "epoch": 0.5671679041617271, + "grad_norm": 4.0501861572265625, + "learning_rate": 0.00016218984630103368, + "loss": 0.1399, + "step": 36360 + }, + { + "epoch": 0.5673238909340488, + "grad_norm": 0.014400321058928967, + "learning_rate": 0.0001621794471828789, + "loss": 0.2372, + "step": 36370 + }, + { + "epoch": 0.5674798777063705, + "grad_norm": 0.4061933755874634, + "learning_rate": 0.00016216904806472412, + "loss": 0.3854, + "step": 36380 + }, + { + "epoch": 0.5676358644786922, + "grad_norm": 0.4940955340862274, + "learning_rate": 0.00016215864894656933, + "loss": 0.1689, + "step": 36390 + }, + { + "epoch": 0.567791851251014, + "grad_norm": 0.3220517635345459, + "learning_rate": 0.00016214824982841455, + "loss": 0.1912, + "step": 36400 + }, + { + "epoch": 0.5679478380233356, + "grad_norm": 2.0848920345306396, + "learning_rate": 0.00016213785071025977, + "loss": 0.1639, + "step": 36410 + }, + { + "epoch": 0.5681038247956574, + "grad_norm": 0.04129406809806824, + "learning_rate": 0.000162127451592105, + "loss": 0.1378, + "step": 36420 + }, + { + "epoch": 0.568259811567979, + "grad_norm": 1.9343401193618774, + "learning_rate": 0.0001621170524739502, + "loss": 0.3759, + "step": 36430 + }, + { + "epoch": 0.5684157983403008, + "grad_norm": 1.0625615119934082, + "learning_rate": 0.00016210665335579543, + "loss": 0.2074, + "step": 36440 + }, + { + "epoch": 0.5685717851126224, + "grad_norm": 0.14784491062164307, + "learning_rate": 0.00016209625423764065, + "loss": 0.0981, + "step": 36450 + }, + { + "epoch": 0.5687277718849442, + "grad_norm": 1.8117821216583252, + "learning_rate": 0.00016208585511948587, + "loss": 0.2755, + "step": 36460 + }, + { + "epoch": 0.5688837586572658, + "grad_norm": 0.061010006815195084, + "learning_rate": 0.0001620754560013311, + "loss": 0.1815, + "step": 36470 + }, + { + "epoch": 0.5690397454295876, + "grad_norm": 0.44453224539756775, + "learning_rate": 0.0001620650568831763, + "loss": 0.0891, + "step": 36480 + }, + { + "epoch": 0.5691957322019093, + "grad_norm": 0.6130645871162415, + "learning_rate": 0.00016205465776502153, + "loss": 0.2778, + "step": 36490 + }, + { + "epoch": 0.569351718974231, + "grad_norm": 5.408796787261963, + "learning_rate": 0.00016204425864686678, + "loss": 0.2301, + "step": 36500 + }, + { + "epoch": 0.5695077057465527, + "grad_norm": 1.2095412015914917, + "learning_rate": 0.00016203385952871197, + "loss": 0.1378, + "step": 36510 + }, + { + "epoch": 0.5696636925188744, + "grad_norm": 0.46873530745506287, + "learning_rate": 0.0001620234604105572, + "loss": 0.1431, + "step": 36520 + }, + { + "epoch": 0.5698196792911961, + "grad_norm": 1.028438925743103, + "learning_rate": 0.0001620130612924024, + "loss": 0.1114, + "step": 36530 + }, + { + "epoch": 0.5699756660635178, + "grad_norm": 0.5952587127685547, + "learning_rate": 0.00016200266217424763, + "loss": 0.2156, + "step": 36540 + }, + { + "epoch": 0.5701316528358396, + "grad_norm": 0.5134285688400269, + "learning_rate": 0.00016199226305609285, + "loss": 0.2712, + "step": 36550 + }, + { + "epoch": 0.5702876396081612, + "grad_norm": 0.03440163657069206, + "learning_rate": 0.00016198186393793807, + "loss": 0.312, + "step": 36560 + }, + { + "epoch": 0.570443626380483, + "grad_norm": 0.10363951325416565, + "learning_rate": 0.0001619714648197833, + "loss": 0.2399, + "step": 36570 + }, + { + "epoch": 0.5705996131528046, + "grad_norm": 1.3341947793960571, + "learning_rate": 0.0001619610657016285, + "loss": 0.1397, + "step": 36580 + }, + { + "epoch": 0.5707555999251264, + "grad_norm": 0.0668073296546936, + "learning_rate": 0.00016195066658347373, + "loss": 0.1091, + "step": 36590 + }, + { + "epoch": 0.570911586697448, + "grad_norm": 5.0486674308776855, + "learning_rate": 0.00016194026746531895, + "loss": 0.4893, + "step": 36600 + }, + { + "epoch": 0.5710675734697698, + "grad_norm": 1.2655692100524902, + "learning_rate": 0.00016192986834716417, + "loss": 0.3344, + "step": 36610 + }, + { + "epoch": 0.5712235602420914, + "grad_norm": 0.1458190381526947, + "learning_rate": 0.00016191946922900939, + "loss": 0.247, + "step": 36620 + }, + { + "epoch": 0.5713795470144132, + "grad_norm": 2.5789072513580322, + "learning_rate": 0.0001619090701108546, + "loss": 0.1546, + "step": 36630 + }, + { + "epoch": 0.5715355337867349, + "grad_norm": 1.191821575164795, + "learning_rate": 0.00016189867099269983, + "loss": 0.2215, + "step": 36640 + }, + { + "epoch": 0.5716915205590566, + "grad_norm": 8.163617134094238, + "learning_rate": 0.00016188827187454505, + "loss": 0.1696, + "step": 36650 + }, + { + "epoch": 0.5718475073313783, + "grad_norm": 3.162277936935425, + "learning_rate": 0.00016187787275639027, + "loss": 0.2906, + "step": 36660 + }, + { + "epoch": 0.5720034941037, + "grad_norm": 2.2368602752685547, + "learning_rate": 0.00016186747363823548, + "loss": 0.3806, + "step": 36670 + }, + { + "epoch": 0.5721594808760218, + "grad_norm": 2.0193803310394287, + "learning_rate": 0.0001618570745200807, + "loss": 0.1964, + "step": 36680 + }, + { + "epoch": 0.5723154676483434, + "grad_norm": 0.9131811261177063, + "learning_rate": 0.00016184667540192592, + "loss": 0.1902, + "step": 36690 + }, + { + "epoch": 0.5724714544206652, + "grad_norm": 0.21330870687961578, + "learning_rate": 0.00016183627628377114, + "loss": 0.2907, + "step": 36700 + }, + { + "epoch": 0.5726274411929868, + "grad_norm": 0.7145895957946777, + "learning_rate": 0.00016182587716561636, + "loss": 0.2574, + "step": 36710 + }, + { + "epoch": 0.5727834279653086, + "grad_norm": 1.0699433088302612, + "learning_rate": 0.00016181547804746158, + "loss": 0.1675, + "step": 36720 + }, + { + "epoch": 0.5729394147376302, + "grad_norm": 0.5066679120063782, + "learning_rate": 0.0001618050789293068, + "loss": 0.4487, + "step": 36730 + }, + { + "epoch": 0.573095401509952, + "grad_norm": 0.7482333183288574, + "learning_rate": 0.00016179467981115202, + "loss": 0.4101, + "step": 36740 + }, + { + "epoch": 0.5732513882822736, + "grad_norm": 0.15925532579421997, + "learning_rate": 0.00016178428069299724, + "loss": 0.2825, + "step": 36750 + }, + { + "epoch": 0.5734073750545954, + "grad_norm": 0.95689457654953, + "learning_rate": 0.00016177388157484246, + "loss": 0.2754, + "step": 36760 + }, + { + "epoch": 0.573563361826917, + "grad_norm": 2.15043568611145, + "learning_rate": 0.00016176348245668768, + "loss": 0.2073, + "step": 36770 + }, + { + "epoch": 0.5737193485992388, + "grad_norm": 2.651520252227783, + "learning_rate": 0.0001617530833385329, + "loss": 0.2517, + "step": 36780 + }, + { + "epoch": 0.5738753353715605, + "grad_norm": 2.1156702041625977, + "learning_rate": 0.00016174268422037812, + "loss": 0.3029, + "step": 36790 + }, + { + "epoch": 0.5740313221438822, + "grad_norm": 3.6200709342956543, + "learning_rate": 0.00016173228510222334, + "loss": 0.4174, + "step": 36800 + }, + { + "epoch": 0.574187308916204, + "grad_norm": 1.447937250137329, + "learning_rate": 0.00016172188598406856, + "loss": 0.1763, + "step": 36810 + }, + { + "epoch": 0.5743432956885256, + "grad_norm": 1.4020819664001465, + "learning_rate": 0.00016171148686591378, + "loss": 0.3994, + "step": 36820 + }, + { + "epoch": 0.5744992824608474, + "grad_norm": 1.4079807996749878, + "learning_rate": 0.000161701087747759, + "loss": 0.2051, + "step": 36830 + }, + { + "epoch": 0.574655269233169, + "grad_norm": 0.05268567427992821, + "learning_rate": 0.00016169068862960422, + "loss": 0.1773, + "step": 36840 + }, + { + "epoch": 0.5748112560054908, + "grad_norm": 2.248767614364624, + "learning_rate": 0.00016168028951144944, + "loss": 0.2534, + "step": 36850 + }, + { + "epoch": 0.5749672427778124, + "grad_norm": 0.3726302981376648, + "learning_rate": 0.00016166989039329466, + "loss": 0.328, + "step": 36860 + }, + { + "epoch": 0.5751232295501342, + "grad_norm": 1.0772110223770142, + "learning_rate": 0.00016165949127513988, + "loss": 0.2069, + "step": 36870 + }, + { + "epoch": 0.5752792163224558, + "grad_norm": 0.11339398473501205, + "learning_rate": 0.0001616490921569851, + "loss": 0.1827, + "step": 36880 + }, + { + "epoch": 0.5754352030947776, + "grad_norm": 1.2981103658676147, + "learning_rate": 0.00016163869303883032, + "loss": 0.1378, + "step": 36890 + }, + { + "epoch": 0.5755911898670992, + "grad_norm": 0.3302285075187683, + "learning_rate": 0.00016162829392067554, + "loss": 0.2311, + "step": 36900 + }, + { + "epoch": 0.575747176639421, + "grad_norm": 0.10168436169624329, + "learning_rate": 0.00016161789480252076, + "loss": 0.1678, + "step": 36910 + }, + { + "epoch": 0.5759031634117426, + "grad_norm": 1.6778357028961182, + "learning_rate": 0.00016160749568436598, + "loss": 0.2385, + "step": 36920 + }, + { + "epoch": 0.5760591501840644, + "grad_norm": 8.1710205078125, + "learning_rate": 0.0001615970965662112, + "loss": 0.3287, + "step": 36930 + }, + { + "epoch": 0.5762151369563862, + "grad_norm": 2.6481053829193115, + "learning_rate": 0.00016158669744805641, + "loss": 0.3119, + "step": 36940 + }, + { + "epoch": 0.5763711237287078, + "grad_norm": 0.3120105266571045, + "learning_rate": 0.00016157629832990163, + "loss": 0.1985, + "step": 36950 + }, + { + "epoch": 0.5765271105010296, + "grad_norm": 2.7221410274505615, + "learning_rate": 0.00016156589921174685, + "loss": 0.3697, + "step": 36960 + }, + { + "epoch": 0.5766830972733512, + "grad_norm": 0.34397706389427185, + "learning_rate": 0.00016155550009359207, + "loss": 0.2162, + "step": 36970 + }, + { + "epoch": 0.576839084045673, + "grad_norm": 4.6421003341674805, + "learning_rate": 0.0001615451009754373, + "loss": 0.2892, + "step": 36980 + }, + { + "epoch": 0.5769950708179946, + "grad_norm": 1.5395855903625488, + "learning_rate": 0.0001615347018572825, + "loss": 0.1258, + "step": 36990 + }, + { + "epoch": 0.5771510575903164, + "grad_norm": 0.4586782455444336, + "learning_rate": 0.00016152430273912773, + "loss": 0.1752, + "step": 37000 + }, + { + "epoch": 0.577307044362638, + "grad_norm": 0.4277658760547638, + "learning_rate": 0.00016151390362097295, + "loss": 0.1722, + "step": 37010 + }, + { + "epoch": 0.5774630311349598, + "grad_norm": 1.0282249450683594, + "learning_rate": 0.00016150350450281817, + "loss": 0.1361, + "step": 37020 + }, + { + "epoch": 0.5776190179072814, + "grad_norm": 0.2496921867132187, + "learning_rate": 0.0001614931053846634, + "loss": 0.0954, + "step": 37030 + }, + { + "epoch": 0.5777750046796032, + "grad_norm": 1.7900562286376953, + "learning_rate": 0.0001614827062665086, + "loss": 0.1284, + "step": 37040 + }, + { + "epoch": 0.5779309914519248, + "grad_norm": 0.6629091501235962, + "learning_rate": 0.00016147230714835383, + "loss": 0.2657, + "step": 37050 + }, + { + "epoch": 0.5780869782242466, + "grad_norm": 2.502882242202759, + "learning_rate": 0.00016146190803019905, + "loss": 0.207, + "step": 37060 + }, + { + "epoch": 0.5782429649965682, + "grad_norm": 1.486069917678833, + "learning_rate": 0.00016145150891204427, + "loss": 0.1232, + "step": 37070 + }, + { + "epoch": 0.57839895176889, + "grad_norm": 0.13543163239955902, + "learning_rate": 0.0001614411097938895, + "loss": 0.3245, + "step": 37080 + }, + { + "epoch": 0.5785549385412118, + "grad_norm": 3.9703898429870605, + "learning_rate": 0.0001614307106757347, + "loss": 0.3104, + "step": 37090 + }, + { + "epoch": 0.5787109253135334, + "grad_norm": 3.2236313819885254, + "learning_rate": 0.00016142031155757993, + "loss": 0.1739, + "step": 37100 + }, + { + "epoch": 0.5788669120858552, + "grad_norm": 0.3854866325855255, + "learning_rate": 0.00016140991243942515, + "loss": 0.084, + "step": 37110 + }, + { + "epoch": 0.5790228988581768, + "grad_norm": 0.1981169581413269, + "learning_rate": 0.00016139951332127037, + "loss": 0.2367, + "step": 37120 + }, + { + "epoch": 0.5791788856304986, + "grad_norm": 2.7315785884857178, + "learning_rate": 0.0001613891142031156, + "loss": 0.1816, + "step": 37130 + }, + { + "epoch": 0.5793348724028202, + "grad_norm": 0.48251378536224365, + "learning_rate": 0.0001613787150849608, + "loss": 0.2123, + "step": 37140 + }, + { + "epoch": 0.579490859175142, + "grad_norm": 5.0431413650512695, + "learning_rate": 0.00016136831596680603, + "loss": 0.2489, + "step": 37150 + }, + { + "epoch": 0.5796468459474636, + "grad_norm": 0.81981360912323, + "learning_rate": 0.00016135791684865125, + "loss": 0.5235, + "step": 37160 + }, + { + "epoch": 0.5798028327197854, + "grad_norm": 0.5701965689659119, + "learning_rate": 0.00016134751773049647, + "loss": 0.1709, + "step": 37170 + }, + { + "epoch": 0.579958819492107, + "grad_norm": 1.3905616998672485, + "learning_rate": 0.00016133711861234169, + "loss": 0.2641, + "step": 37180 + }, + { + "epoch": 0.5801148062644288, + "grad_norm": 2.281285047531128, + "learning_rate": 0.0001613267194941869, + "loss": 0.3915, + "step": 37190 + }, + { + "epoch": 0.5802707930367504, + "grad_norm": 1.2038309574127197, + "learning_rate": 0.00016131632037603213, + "loss": 0.0881, + "step": 37200 + }, + { + "epoch": 0.5804267798090722, + "grad_norm": 0.3096400201320648, + "learning_rate": 0.00016130592125787735, + "loss": 0.1341, + "step": 37210 + }, + { + "epoch": 0.5805827665813938, + "grad_norm": 0.10219905525445938, + "learning_rate": 0.00016129552213972256, + "loss": 0.0867, + "step": 37220 + }, + { + "epoch": 0.5807387533537156, + "grad_norm": 0.13869646191596985, + "learning_rate": 0.00016128512302156778, + "loss": 0.1621, + "step": 37230 + }, + { + "epoch": 0.5808947401260373, + "grad_norm": 0.7992938160896301, + "learning_rate": 0.000161274723903413, + "loss": 0.275, + "step": 37240 + }, + { + "epoch": 0.581050726898359, + "grad_norm": 0.2635972201824188, + "learning_rate": 0.00016126432478525822, + "loss": 0.2429, + "step": 37250 + }, + { + "epoch": 0.5812067136706808, + "grad_norm": 1.7211238145828247, + "learning_rate": 0.00016125392566710344, + "loss": 0.1541, + "step": 37260 + }, + { + "epoch": 0.5813627004430024, + "grad_norm": 2.2448506355285645, + "learning_rate": 0.00016124352654894866, + "loss": 0.3689, + "step": 37270 + }, + { + "epoch": 0.5815186872153242, + "grad_norm": 1.238708257675171, + "learning_rate": 0.00016123312743079388, + "loss": 0.3292, + "step": 37280 + }, + { + "epoch": 0.5816746739876458, + "grad_norm": 0.5670503377914429, + "learning_rate": 0.0001612227283126391, + "loss": 0.1151, + "step": 37290 + }, + { + "epoch": 0.5818306607599676, + "grad_norm": 2.287304401397705, + "learning_rate": 0.00016121232919448432, + "loss": 0.3558, + "step": 37300 + }, + { + "epoch": 0.5819866475322892, + "grad_norm": 0.8009629845619202, + "learning_rate": 0.00016120193007632954, + "loss": 0.1861, + "step": 37310 + }, + { + "epoch": 0.582142634304611, + "grad_norm": 1.1901466846466064, + "learning_rate": 0.00016119153095817476, + "loss": 0.089, + "step": 37320 + }, + { + "epoch": 0.5822986210769326, + "grad_norm": 1.7198141813278198, + "learning_rate": 0.00016118113184001998, + "loss": 0.2931, + "step": 37330 + }, + { + "epoch": 0.5824546078492544, + "grad_norm": 0.8832749724388123, + "learning_rate": 0.0001611707327218652, + "loss": 0.2342, + "step": 37340 + }, + { + "epoch": 0.582610594621576, + "grad_norm": 1.0623290538787842, + "learning_rate": 0.00016116033360371042, + "loss": 0.2317, + "step": 37350 + }, + { + "epoch": 0.5827665813938978, + "grad_norm": 0.5173699855804443, + "learning_rate": 0.00016114993448555564, + "loss": 0.1299, + "step": 37360 + }, + { + "epoch": 0.5829225681662195, + "grad_norm": 0.05118720978498459, + "learning_rate": 0.00016113953536740086, + "loss": 0.0786, + "step": 37370 + }, + { + "epoch": 0.5830785549385412, + "grad_norm": 2.0060107707977295, + "learning_rate": 0.00016112913624924608, + "loss": 0.266, + "step": 37380 + }, + { + "epoch": 0.5832345417108629, + "grad_norm": 5.121693134307861, + "learning_rate": 0.0001611187371310913, + "loss": 0.1824, + "step": 37390 + }, + { + "epoch": 0.5833905284831846, + "grad_norm": 0.3568089008331299, + "learning_rate": 0.00016110833801293652, + "loss": 0.3189, + "step": 37400 + }, + { + "epoch": 0.5835465152555064, + "grad_norm": 0.8631492853164673, + "learning_rate": 0.00016109793889478174, + "loss": 0.1724, + "step": 37410 + }, + { + "epoch": 0.583702502027828, + "grad_norm": 1.9214404821395874, + "learning_rate": 0.00016108753977662693, + "loss": 0.2094, + "step": 37420 + }, + { + "epoch": 0.5838584888001498, + "grad_norm": 1.9055646657943726, + "learning_rate": 0.00016107714065847218, + "loss": 0.1258, + "step": 37430 + }, + { + "epoch": 0.5840144755724714, + "grad_norm": 0.6614194512367249, + "learning_rate": 0.00016106674154031737, + "loss": 0.1966, + "step": 37440 + }, + { + "epoch": 0.5841704623447932, + "grad_norm": 0.286883682012558, + "learning_rate": 0.00016105634242216262, + "loss": 0.588, + "step": 37450 + }, + { + "epoch": 0.5843264491171148, + "grad_norm": 0.5599405765533447, + "learning_rate": 0.0001610459433040078, + "loss": 0.3414, + "step": 37460 + }, + { + "epoch": 0.5844824358894366, + "grad_norm": 1.5908915996551514, + "learning_rate": 0.00016103554418585306, + "loss": 0.1264, + "step": 37470 + }, + { + "epoch": 0.5846384226617583, + "grad_norm": 1.0729930400848389, + "learning_rate": 0.00016102514506769825, + "loss": 0.4815, + "step": 37480 + }, + { + "epoch": 0.58479440943408, + "grad_norm": 0.8912356495857239, + "learning_rate": 0.0001610147459495435, + "loss": 0.2092, + "step": 37490 + }, + { + "epoch": 0.5849503962064017, + "grad_norm": 0.24116197228431702, + "learning_rate": 0.0001610043468313887, + "loss": 0.1552, + "step": 37500 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.2835920896e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-37500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..33b4634925ce50209dc8e4c13c5ef3a6fd7c0c31 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:099b99a5b08e03b3ab27fd46ec563c88402f8ec61514288693965c4433c249f6 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..2faf30707f9f0c1c4c3fa617c6e0186173858fdf --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2dce146c3b3aa1c0f76d7622a17a90968130c87f202674cdbdaf81c10c818d21 +size 66060288 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..fbc52f8126b66c6dc0a628c98b48f99d02110361 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..84fee5ebf7dcb2c18310c4ab67452a25da9cc4ae Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..13a03da185fe29ce40113c03facac8676ae59d7e --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/trainer_state.json @@ -0,0 +1,26634 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.592749734822487, + "eval_steps": 500, + "global_step": 38000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + }, + { + "epoch": 0.5071129968178698, + "grad_norm": 1.0075199604034424, + "learning_rate": 0.00016619350679062414, + "loss": 0.2569, + "step": 32510 + }, + { + "epoch": 0.5072689835901916, + "grad_norm": 1.3489066362380981, + "learning_rate": 0.0001661831076724694, + "loss": 0.2565, + "step": 32520 + }, + { + "epoch": 0.5074249703625132, + "grad_norm": 0.2534058094024658, + "learning_rate": 0.00016617270855431458, + "loss": 0.2292, + "step": 32530 + }, + { + "epoch": 0.507580957134835, + "grad_norm": 0.31723281741142273, + "learning_rate": 0.00016616230943615983, + "loss": 0.1467, + "step": 32540 + }, + { + "epoch": 0.5077369439071566, + "grad_norm": 0.29659849405288696, + "learning_rate": 0.00016615191031800502, + "loss": 0.2768, + "step": 32550 + }, + { + "epoch": 0.5078929306794784, + "grad_norm": 1.5246435403823853, + "learning_rate": 0.00016614151119985027, + "loss": 0.216, + "step": 32560 + }, + { + "epoch": 0.5080489174518001, + "grad_norm": 0.6026854515075684, + "learning_rate": 0.00016613111208169546, + "loss": 0.2277, + "step": 32570 + }, + { + "epoch": 0.5082049042241218, + "grad_norm": 0.41994890570640564, + "learning_rate": 0.0001661207129635407, + "loss": 0.2433, + "step": 32580 + }, + { + "epoch": 0.5083608909964435, + "grad_norm": 2.084599733352661, + "learning_rate": 0.00016611031384538593, + "loss": 0.4081, + "step": 32590 + }, + { + "epoch": 0.5085168777687652, + "grad_norm": 0.2945229709148407, + "learning_rate": 0.00016609991472723115, + "loss": 0.1329, + "step": 32600 + }, + { + "epoch": 0.508672864541087, + "grad_norm": 1.4981423616409302, + "learning_rate": 0.00016608951560907636, + "loss": 0.2855, + "step": 32610 + }, + { + "epoch": 0.5088288513134086, + "grad_norm": 0.961737871170044, + "learning_rate": 0.00016607911649092158, + "loss": 0.1654, + "step": 32620 + }, + { + "epoch": 0.5089848380857304, + "grad_norm": 1.49734628200531, + "learning_rate": 0.0001660687173727668, + "loss": 0.2081, + "step": 32630 + }, + { + "epoch": 0.509140824858052, + "grad_norm": 2.538135051727295, + "learning_rate": 0.00016605831825461202, + "loss": 0.3023, + "step": 32640 + }, + { + "epoch": 0.5092968116303738, + "grad_norm": 0.2981272041797638, + "learning_rate": 0.00016604791913645724, + "loss": 0.214, + "step": 32650 + }, + { + "epoch": 0.5094527984026954, + "grad_norm": 1.8863650560379028, + "learning_rate": 0.00016603752001830246, + "loss": 0.3363, + "step": 32660 + }, + { + "epoch": 0.5096087851750172, + "grad_norm": 0.022720983251929283, + "learning_rate": 0.00016602712090014768, + "loss": 0.2463, + "step": 32670 + }, + { + "epoch": 0.5097647719473388, + "grad_norm": 1.154977798461914, + "learning_rate": 0.0001660167217819929, + "loss": 0.4484, + "step": 32680 + }, + { + "epoch": 0.5099207587196606, + "grad_norm": 0.9470499157905579, + "learning_rate": 0.00016600632266383812, + "loss": 0.1196, + "step": 32690 + }, + { + "epoch": 0.5100767454919822, + "grad_norm": 0.19955401122570038, + "learning_rate": 0.00016599592354568334, + "loss": 0.1762, + "step": 32700 + }, + { + "epoch": 0.510232732264304, + "grad_norm": 0.4098282754421234, + "learning_rate": 0.00016598552442752856, + "loss": 0.2739, + "step": 32710 + }, + { + "epoch": 0.5103887190366257, + "grad_norm": 0.21697258949279785, + "learning_rate": 0.00016597512530937378, + "loss": 0.2555, + "step": 32720 + }, + { + "epoch": 0.5105447058089474, + "grad_norm": 4.894479751586914, + "learning_rate": 0.000165964726191219, + "loss": 0.3137, + "step": 32730 + }, + { + "epoch": 0.5107006925812692, + "grad_norm": 0.8396711349487305, + "learning_rate": 0.00016595432707306422, + "loss": 0.3105, + "step": 32740 + }, + { + "epoch": 0.5108566793535908, + "grad_norm": 1.5494674444198608, + "learning_rate": 0.00016594392795490944, + "loss": 0.2271, + "step": 32750 + }, + { + "epoch": 0.5110126661259126, + "grad_norm": 1.3409888744354248, + "learning_rate": 0.00016593352883675466, + "loss": 0.1796, + "step": 32760 + }, + { + "epoch": 0.5111686528982342, + "grad_norm": 1.7590911388397217, + "learning_rate": 0.00016592312971859988, + "loss": 0.2185, + "step": 32770 + }, + { + "epoch": 0.511324639670556, + "grad_norm": 0.7350002527236938, + "learning_rate": 0.0001659127306004451, + "loss": 0.1911, + "step": 32780 + }, + { + "epoch": 0.5114806264428776, + "grad_norm": 0.7176411747932434, + "learning_rate": 0.00016590233148229032, + "loss": 0.6918, + "step": 32790 + }, + { + "epoch": 0.5116366132151994, + "grad_norm": 2.5005321502685547, + "learning_rate": 0.00016589193236413554, + "loss": 0.2329, + "step": 32800 + }, + { + "epoch": 0.511792599987521, + "grad_norm": 1.3916250467300415, + "learning_rate": 0.00016588153324598076, + "loss": 0.2102, + "step": 32810 + }, + { + "epoch": 0.5119485867598428, + "grad_norm": 0.721671998500824, + "learning_rate": 0.00016587113412782598, + "loss": 0.2965, + "step": 32820 + }, + { + "epoch": 0.5121045735321644, + "grad_norm": 0.6120114326477051, + "learning_rate": 0.0001658607350096712, + "loss": 0.2661, + "step": 32830 + }, + { + "epoch": 0.5122605603044862, + "grad_norm": 0.29123347997665405, + "learning_rate": 0.00016585033589151642, + "loss": 0.2455, + "step": 32840 + }, + { + "epoch": 0.5124165470768078, + "grad_norm": 2.7300868034362793, + "learning_rate": 0.00016583993677336164, + "loss": 0.2049, + "step": 32850 + }, + { + "epoch": 0.5125725338491296, + "grad_norm": 1.801928997039795, + "learning_rate": 0.00016582953765520686, + "loss": 0.1725, + "step": 32860 + }, + { + "epoch": 0.5127285206214514, + "grad_norm": 0.6406471729278564, + "learning_rate": 0.00016581913853705208, + "loss": 0.164, + "step": 32870 + }, + { + "epoch": 0.512884507393773, + "grad_norm": 0.9627325534820557, + "learning_rate": 0.0001658087394188973, + "loss": 0.1728, + "step": 32880 + }, + { + "epoch": 0.5130404941660948, + "grad_norm": 1.5446962118148804, + "learning_rate": 0.00016579834030074251, + "loss": 0.2345, + "step": 32890 + }, + { + "epoch": 0.5131964809384164, + "grad_norm": 0.28092706203460693, + "learning_rate": 0.0001657879411825877, + "loss": 0.2124, + "step": 32900 + }, + { + "epoch": 0.5133524677107382, + "grad_norm": 1.7751421928405762, + "learning_rate": 0.00016577754206443295, + "loss": 0.2367, + "step": 32910 + }, + { + "epoch": 0.5135084544830598, + "grad_norm": 1.6964634656906128, + "learning_rate": 0.00016576714294627815, + "loss": 0.2361, + "step": 32920 + }, + { + "epoch": 0.5136644412553816, + "grad_norm": 0.06759823113679886, + "learning_rate": 0.0001657567438281234, + "loss": 0.2329, + "step": 32930 + }, + { + "epoch": 0.5138204280277032, + "grad_norm": 2.2519280910491943, + "learning_rate": 0.00016574634470996859, + "loss": 0.1079, + "step": 32940 + }, + { + "epoch": 0.513976414800025, + "grad_norm": 3.167576551437378, + "learning_rate": 0.00016573594559181383, + "loss": 0.1764, + "step": 32950 + }, + { + "epoch": 0.5141324015723466, + "grad_norm": 1.6734397411346436, + "learning_rate": 0.00016572554647365903, + "loss": 0.2878, + "step": 32960 + }, + { + "epoch": 0.5142883883446684, + "grad_norm": 0.35627374053001404, + "learning_rate": 0.00016571514735550427, + "loss": 0.2337, + "step": 32970 + }, + { + "epoch": 0.51444437511699, + "grad_norm": 1.6321039199829102, + "learning_rate": 0.00016570474823734946, + "loss": 0.2551, + "step": 32980 + }, + { + "epoch": 0.5146003618893118, + "grad_norm": 1.352982997894287, + "learning_rate": 0.0001656943491191947, + "loss": 0.4467, + "step": 32990 + }, + { + "epoch": 0.5147563486616334, + "grad_norm": 0.3558129072189331, + "learning_rate": 0.0001656839500010399, + "loss": 0.1913, + "step": 33000 + }, + { + "epoch": 0.5149123354339552, + "grad_norm": 0.041972286999225616, + "learning_rate": 0.00016567355088288515, + "loss": 0.1704, + "step": 33010 + }, + { + "epoch": 0.515068322206277, + "grad_norm": 1.4168939590454102, + "learning_rate": 0.00016566315176473034, + "loss": 0.4866, + "step": 33020 + }, + { + "epoch": 0.5152243089785986, + "grad_norm": 3.0042929649353027, + "learning_rate": 0.0001656527526465756, + "loss": 0.2473, + "step": 33030 + }, + { + "epoch": 0.5153802957509204, + "grad_norm": 1.027274489402771, + "learning_rate": 0.00016564235352842078, + "loss": 0.436, + "step": 33040 + }, + { + "epoch": 0.515536282523242, + "grad_norm": 1.2330195903778076, + "learning_rate": 0.00016563195441026603, + "loss": 0.4615, + "step": 33050 + }, + { + "epoch": 0.5156922692955638, + "grad_norm": 2.7145066261291504, + "learning_rate": 0.00016562155529211122, + "loss": 0.278, + "step": 33060 + }, + { + "epoch": 0.5158482560678854, + "grad_norm": 1.1273893117904663, + "learning_rate": 0.00016561115617395647, + "loss": 0.3767, + "step": 33070 + }, + { + "epoch": 0.5160042428402072, + "grad_norm": 2.466265916824341, + "learning_rate": 0.00016560075705580166, + "loss": 0.1879, + "step": 33080 + }, + { + "epoch": 0.5161602296125288, + "grad_norm": 0.3759515583515167, + "learning_rate": 0.0001655903579376469, + "loss": 0.4004, + "step": 33090 + }, + { + "epoch": 0.5163162163848506, + "grad_norm": 1.8821336030960083, + "learning_rate": 0.0001655799588194921, + "loss": 0.1622, + "step": 33100 + }, + { + "epoch": 0.5164722031571722, + "grad_norm": 0.8162096738815308, + "learning_rate": 0.00016556955970133735, + "loss": 0.1341, + "step": 33110 + }, + { + "epoch": 0.516628189929494, + "grad_norm": 0.008498801849782467, + "learning_rate": 0.00016555916058318254, + "loss": 0.2546, + "step": 33120 + }, + { + "epoch": 0.5167841767018156, + "grad_norm": 0.17536675930023193, + "learning_rate": 0.00016554876146502779, + "loss": 0.2934, + "step": 33130 + }, + { + "epoch": 0.5169401634741374, + "grad_norm": 3.539898633956909, + "learning_rate": 0.00016553836234687298, + "loss": 0.4959, + "step": 33140 + }, + { + "epoch": 0.517096150246459, + "grad_norm": 2.190887212753296, + "learning_rate": 0.00016552796322871823, + "loss": 0.2914, + "step": 33150 + }, + { + "epoch": 0.5172521370187808, + "grad_norm": 1.6388943195343018, + "learning_rate": 0.00016551756411056342, + "loss": 0.1317, + "step": 33160 + }, + { + "epoch": 0.5174081237911026, + "grad_norm": 0.33063387870788574, + "learning_rate": 0.00016550716499240866, + "loss": 0.4258, + "step": 33170 + }, + { + "epoch": 0.5175641105634242, + "grad_norm": 0.5168288946151733, + "learning_rate": 0.00016549676587425386, + "loss": 0.2548, + "step": 33180 + }, + { + "epoch": 0.517720097335746, + "grad_norm": 3.9570281505584717, + "learning_rate": 0.0001654863667560991, + "loss": 0.3179, + "step": 33190 + }, + { + "epoch": 0.5178760841080676, + "grad_norm": 1.127255916595459, + "learning_rate": 0.0001654759676379443, + "loss": 0.233, + "step": 33200 + }, + { + "epoch": 0.5180320708803894, + "grad_norm": 0.36299410462379456, + "learning_rate": 0.00016546556851978954, + "loss": 0.4614, + "step": 33210 + }, + { + "epoch": 0.518188057652711, + "grad_norm": 2.914668321609497, + "learning_rate": 0.00016545516940163474, + "loss": 0.1603, + "step": 33220 + }, + { + "epoch": 0.5183440444250328, + "grad_norm": 6.991094589233398, + "learning_rate": 0.00016544477028347998, + "loss": 0.2259, + "step": 33230 + }, + { + "epoch": 0.5185000311973544, + "grad_norm": 2.0594241619110107, + "learning_rate": 0.00016543437116532517, + "loss": 0.1885, + "step": 33240 + }, + { + "epoch": 0.5186560179696762, + "grad_norm": 0.5594693422317505, + "learning_rate": 0.00016542397204717042, + "loss": 0.4333, + "step": 33250 + }, + { + "epoch": 0.5188120047419978, + "grad_norm": 4.314854621887207, + "learning_rate": 0.00016541357292901561, + "loss": 0.2103, + "step": 33260 + }, + { + "epoch": 0.5189679915143196, + "grad_norm": 0.5609000325202942, + "learning_rate": 0.00016540317381086086, + "loss": 0.2498, + "step": 33270 + }, + { + "epoch": 0.5191239782866413, + "grad_norm": 3.680088758468628, + "learning_rate": 0.00016539277469270605, + "loss": 0.3465, + "step": 33280 + }, + { + "epoch": 0.519279965058963, + "grad_norm": 1.4993672370910645, + "learning_rate": 0.0001653823755745513, + "loss": 0.3272, + "step": 33290 + }, + { + "epoch": 0.5194359518312847, + "grad_norm": 0.18005084991455078, + "learning_rate": 0.0001653719764563965, + "loss": 0.1706, + "step": 33300 + }, + { + "epoch": 0.5195919386036064, + "grad_norm": 0.5105524659156799, + "learning_rate": 0.00016536157733824174, + "loss": 0.0937, + "step": 33310 + }, + { + "epoch": 0.5197479253759281, + "grad_norm": 1.2806589603424072, + "learning_rate": 0.00016535117822008693, + "loss": 0.1578, + "step": 33320 + }, + { + "epoch": 0.5199039121482498, + "grad_norm": 0.22015990316867828, + "learning_rate": 0.00016534077910193218, + "loss": 0.2081, + "step": 33330 + }, + { + "epoch": 0.5200598989205716, + "grad_norm": 0.34638991951942444, + "learning_rate": 0.00016533037998377737, + "loss": 0.2389, + "step": 33340 + }, + { + "epoch": 0.5202158856928932, + "grad_norm": 0.8709302544593811, + "learning_rate": 0.0001653199808656226, + "loss": 0.2112, + "step": 33350 + }, + { + "epoch": 0.520371872465215, + "grad_norm": 1.2309521436691284, + "learning_rate": 0.0001653095817474678, + "loss": 0.3448, + "step": 33360 + }, + { + "epoch": 0.5205278592375366, + "grad_norm": 1.3155115842819214, + "learning_rate": 0.00016529918262931303, + "loss": 0.1572, + "step": 33370 + }, + { + "epoch": 0.5206838460098584, + "grad_norm": 0.358751118183136, + "learning_rate": 0.00016528878351115825, + "loss": 0.3011, + "step": 33380 + }, + { + "epoch": 0.52083983278218, + "grad_norm": 3.800628662109375, + "learning_rate": 0.00016527838439300347, + "loss": 0.458, + "step": 33390 + }, + { + "epoch": 0.5209958195545018, + "grad_norm": 1.1253124475479126, + "learning_rate": 0.0001652679852748487, + "loss": 0.1238, + "step": 33400 + }, + { + "epoch": 0.5211518063268235, + "grad_norm": 0.46979108452796936, + "learning_rate": 0.0001652575861566939, + "loss": 0.0745, + "step": 33410 + }, + { + "epoch": 0.5213077930991452, + "grad_norm": 7.921492576599121, + "learning_rate": 0.00016524718703853913, + "loss": 0.4004, + "step": 33420 + }, + { + "epoch": 0.5214637798714669, + "grad_norm": 1.643127679824829, + "learning_rate": 0.00016523678792038435, + "loss": 0.2858, + "step": 33430 + }, + { + "epoch": 0.5216197666437886, + "grad_norm": 7.096631050109863, + "learning_rate": 0.0001652263888022296, + "loss": 0.4735, + "step": 33440 + }, + { + "epoch": 0.5217757534161103, + "grad_norm": 0.7280219197273254, + "learning_rate": 0.0001652159896840748, + "loss": 0.1841, + "step": 33450 + }, + { + "epoch": 0.521931740188432, + "grad_norm": 0.9415522813796997, + "learning_rate": 0.00016520559056592003, + "loss": 0.2169, + "step": 33460 + }, + { + "epoch": 0.5220877269607537, + "grad_norm": 2.264862060546875, + "learning_rate": 0.00016519519144776523, + "loss": 0.1902, + "step": 33470 + }, + { + "epoch": 0.5222437137330754, + "grad_norm": 2.4566822052001953, + "learning_rate": 0.00016518479232961047, + "loss": 0.3329, + "step": 33480 + }, + { + "epoch": 0.5223997005053972, + "grad_norm": 0.7330378890037537, + "learning_rate": 0.00016517439321145567, + "loss": 0.2341, + "step": 33490 + }, + { + "epoch": 0.5225556872777188, + "grad_norm": 1.6409542560577393, + "learning_rate": 0.0001651639940933009, + "loss": 0.3077, + "step": 33500 + }, + { + "epoch": 0.5227116740500406, + "grad_norm": 1.0647432804107666, + "learning_rate": 0.0001651535949751461, + "loss": 0.1935, + "step": 33510 + }, + { + "epoch": 0.5228676608223622, + "grad_norm": 0.507420539855957, + "learning_rate": 0.00016514319585699135, + "loss": 0.1314, + "step": 33520 + }, + { + "epoch": 0.523023647594684, + "grad_norm": 1.0064164400100708, + "learning_rate": 0.00016513279673883654, + "loss": 0.344, + "step": 33530 + }, + { + "epoch": 0.5231796343670057, + "grad_norm": 1.1936209201812744, + "learning_rate": 0.0001651223976206818, + "loss": 0.1948, + "step": 33540 + }, + { + "epoch": 0.5233356211393274, + "grad_norm": 0.04992926865816116, + "learning_rate": 0.00016511199850252698, + "loss": 0.1363, + "step": 33550 + }, + { + "epoch": 0.5234916079116491, + "grad_norm": 1.8256275653839111, + "learning_rate": 0.00016510159938437223, + "loss": 0.443, + "step": 33560 + }, + { + "epoch": 0.5236475946839708, + "grad_norm": 4.914908409118652, + "learning_rate": 0.00016509120026621742, + "loss": 0.4654, + "step": 33570 + }, + { + "epoch": 0.5238035814562925, + "grad_norm": 1.1625486612319946, + "learning_rate": 0.00016508080114806267, + "loss": 0.1974, + "step": 33580 + }, + { + "epoch": 0.5239595682286142, + "grad_norm": 6.6192522048950195, + "learning_rate": 0.00016507040202990786, + "loss": 0.2131, + "step": 33590 + }, + { + "epoch": 0.5241155550009359, + "grad_norm": 0.7369065880775452, + "learning_rate": 0.0001650600029117531, + "loss": 0.2371, + "step": 33600 + }, + { + "epoch": 0.5242715417732576, + "grad_norm": 1.5238152742385864, + "learning_rate": 0.0001650496037935983, + "loss": 0.1557, + "step": 33610 + }, + { + "epoch": 0.5244275285455793, + "grad_norm": 1.0418007373809814, + "learning_rate": 0.00016503920467544355, + "loss": 0.1878, + "step": 33620 + }, + { + "epoch": 0.524583515317901, + "grad_norm": 0.790117084980011, + "learning_rate": 0.00016502880555728874, + "loss": 0.2195, + "step": 33630 + }, + { + "epoch": 0.5247395020902228, + "grad_norm": 1.6712257862091064, + "learning_rate": 0.000165018406439134, + "loss": 0.1602, + "step": 33640 + }, + { + "epoch": 0.5248954888625444, + "grad_norm": 0.19236230850219727, + "learning_rate": 0.00016500800732097918, + "loss": 0.2526, + "step": 33650 + }, + { + "epoch": 0.5250514756348662, + "grad_norm": 1.3519701957702637, + "learning_rate": 0.00016499760820282443, + "loss": 0.4686, + "step": 33660 + }, + { + "epoch": 0.5252074624071879, + "grad_norm": 1.694342851638794, + "learning_rate": 0.00016498720908466962, + "loss": 0.1859, + "step": 33670 + }, + { + "epoch": 0.5253634491795096, + "grad_norm": 5.225239276885986, + "learning_rate": 0.00016497680996651487, + "loss": 0.2944, + "step": 33680 + }, + { + "epoch": 0.5255194359518313, + "grad_norm": 2.0208842754364014, + "learning_rate": 0.00016496641084836006, + "loss": 0.2421, + "step": 33690 + }, + { + "epoch": 0.525675422724153, + "grad_norm": 0.7954996228218079, + "learning_rate": 0.0001649560117302053, + "loss": 0.2853, + "step": 33700 + }, + { + "epoch": 0.5258314094964747, + "grad_norm": 2.296086072921753, + "learning_rate": 0.0001649456126120505, + "loss": 0.1333, + "step": 33710 + }, + { + "epoch": 0.5259873962687964, + "grad_norm": 1.1779128313064575, + "learning_rate": 0.00016493521349389574, + "loss": 0.1066, + "step": 33720 + }, + { + "epoch": 0.5261433830411181, + "grad_norm": 0.1756065934896469, + "learning_rate": 0.00016492481437574094, + "loss": 0.1352, + "step": 33730 + }, + { + "epoch": 0.5262993698134398, + "grad_norm": 0.13100725412368774, + "learning_rate": 0.00016491441525758618, + "loss": 0.2399, + "step": 33740 + }, + { + "epoch": 0.5264553565857615, + "grad_norm": 5.532008171081543, + "learning_rate": 0.00016490401613943138, + "loss": 0.2896, + "step": 33750 + }, + { + "epoch": 0.5266113433580832, + "grad_norm": 1.319886565208435, + "learning_rate": 0.00016489361702127662, + "loss": 0.3275, + "step": 33760 + }, + { + "epoch": 0.5267673301304049, + "grad_norm": 1.5550974607467651, + "learning_rate": 0.00016488321790312182, + "loss": 0.2677, + "step": 33770 + }, + { + "epoch": 0.5269233169027266, + "grad_norm": 1.8936737775802612, + "learning_rate": 0.00016487281878496706, + "loss": 0.1955, + "step": 33780 + }, + { + "epoch": 0.5270793036750484, + "grad_norm": 0.3653401732444763, + "learning_rate": 0.00016486241966681226, + "loss": 0.0723, + "step": 33790 + }, + { + "epoch": 0.52723529044737, + "grad_norm": 2.861341714859009, + "learning_rate": 0.00016485202054865747, + "loss": 0.2412, + "step": 33800 + }, + { + "epoch": 0.5273912772196918, + "grad_norm": 1.5291428565979004, + "learning_rate": 0.0001648416214305027, + "loss": 0.0871, + "step": 33810 + }, + { + "epoch": 0.5275472639920135, + "grad_norm": 1.0372581481933594, + "learning_rate": 0.00016483122231234791, + "loss": 0.4705, + "step": 33820 + }, + { + "epoch": 0.5277032507643352, + "grad_norm": 1.1943141222000122, + "learning_rate": 0.00016482082319419313, + "loss": 0.2848, + "step": 33830 + }, + { + "epoch": 0.5278592375366569, + "grad_norm": 1.9008225202560425, + "learning_rate": 0.00016481042407603835, + "loss": 0.139, + "step": 33840 + }, + { + "epoch": 0.5280152243089786, + "grad_norm": 2.132089138031006, + "learning_rate": 0.00016480002495788357, + "loss": 0.2119, + "step": 33850 + }, + { + "epoch": 0.5281712110813003, + "grad_norm": 0.24524426460266113, + "learning_rate": 0.0001647896258397288, + "loss": 0.1643, + "step": 33860 + }, + { + "epoch": 0.528327197853622, + "grad_norm": 1.6469637155532837, + "learning_rate": 0.000164779226721574, + "loss": 0.363, + "step": 33870 + }, + { + "epoch": 0.5284831846259437, + "grad_norm": 0.8767328858375549, + "learning_rate": 0.00016476882760341923, + "loss": 0.2632, + "step": 33880 + }, + { + "epoch": 0.5286391713982654, + "grad_norm": 0.06347586214542389, + "learning_rate": 0.00016475842848526445, + "loss": 0.3204, + "step": 33890 + }, + { + "epoch": 0.5287951581705871, + "grad_norm": 0.09782540798187256, + "learning_rate": 0.00016474802936710967, + "loss": 0.2211, + "step": 33900 + }, + { + "epoch": 0.5289511449429088, + "grad_norm": 3.2998859882354736, + "learning_rate": 0.0001647376302489549, + "loss": 0.265, + "step": 33910 + }, + { + "epoch": 0.5291071317152305, + "grad_norm": 0.43594226241111755, + "learning_rate": 0.0001647272311308001, + "loss": 0.194, + "step": 33920 + }, + { + "epoch": 0.5292631184875523, + "grad_norm": 1.5166605710983276, + "learning_rate": 0.00016471683201264533, + "loss": 0.2675, + "step": 33930 + }, + { + "epoch": 0.529419105259874, + "grad_norm": 0.6056640148162842, + "learning_rate": 0.00016470643289449055, + "loss": 0.0576, + "step": 33940 + }, + { + "epoch": 0.5295750920321957, + "grad_norm": 0.25410348176956177, + "learning_rate": 0.00016469603377633577, + "loss": 0.1124, + "step": 33950 + }, + { + "epoch": 0.5297310788045174, + "grad_norm": 1.770642876625061, + "learning_rate": 0.000164685634658181, + "loss": 0.3295, + "step": 33960 + }, + { + "epoch": 0.5298870655768391, + "grad_norm": 0.0607205331325531, + "learning_rate": 0.0001646752355400262, + "loss": 0.2369, + "step": 33970 + }, + { + "epoch": 0.5300430523491608, + "grad_norm": 0.5557095408439636, + "learning_rate": 0.00016466483642187143, + "loss": 0.1681, + "step": 33980 + }, + { + "epoch": 0.5301990391214825, + "grad_norm": 0.5192957520484924, + "learning_rate": 0.00016465443730371665, + "loss": 0.2649, + "step": 33990 + }, + { + "epoch": 0.5303550258938042, + "grad_norm": 0.04804835096001625, + "learning_rate": 0.00016464403818556187, + "loss": 0.2639, + "step": 34000 + }, + { + "epoch": 0.5305110126661259, + "grad_norm": 0.02673129364848137, + "learning_rate": 0.0001646336390674071, + "loss": 0.169, + "step": 34010 + }, + { + "epoch": 0.5306669994384476, + "grad_norm": 1.0084244012832642, + "learning_rate": 0.0001646232399492523, + "loss": 0.353, + "step": 34020 + }, + { + "epoch": 0.5308229862107693, + "grad_norm": 2.2202091217041016, + "learning_rate": 0.00016461284083109753, + "loss": 0.3206, + "step": 34030 + }, + { + "epoch": 0.530978972983091, + "grad_norm": 0.5573744773864746, + "learning_rate": 0.00016460244171294275, + "loss": 0.231, + "step": 34040 + }, + { + "epoch": 0.5311349597554127, + "grad_norm": 0.39700084924697876, + "learning_rate": 0.00016459204259478797, + "loss": 0.1412, + "step": 34050 + }, + { + "epoch": 0.5312909465277345, + "grad_norm": 2.582963228225708, + "learning_rate": 0.00016458164347663319, + "loss": 0.3334, + "step": 34060 + }, + { + "epoch": 0.5314469333000561, + "grad_norm": 0.22781169414520264, + "learning_rate": 0.0001645712443584784, + "loss": 0.2767, + "step": 34070 + }, + { + "epoch": 0.5316029200723779, + "grad_norm": 1.2051042318344116, + "learning_rate": 0.00016456084524032362, + "loss": 0.2045, + "step": 34080 + }, + { + "epoch": 0.5317589068446996, + "grad_norm": 0.42760100960731506, + "learning_rate": 0.00016455044612216884, + "loss": 0.1151, + "step": 34090 + }, + { + "epoch": 0.5319148936170213, + "grad_norm": 0.14440476894378662, + "learning_rate": 0.00016454004700401406, + "loss": 0.2386, + "step": 34100 + }, + { + "epoch": 0.532070880389343, + "grad_norm": 2.2777981758117676, + "learning_rate": 0.00016452964788585928, + "loss": 0.2009, + "step": 34110 + }, + { + "epoch": 0.5322268671616647, + "grad_norm": 0.9206979274749756, + "learning_rate": 0.0001645192487677045, + "loss": 0.2745, + "step": 34120 + }, + { + "epoch": 0.5323828539339864, + "grad_norm": 1.6947574615478516, + "learning_rate": 0.00016450884964954972, + "loss": 0.2584, + "step": 34130 + }, + { + "epoch": 0.5325388407063081, + "grad_norm": 0.401444673538208, + "learning_rate": 0.00016449845053139494, + "loss": 0.2218, + "step": 34140 + }, + { + "epoch": 0.5326948274786298, + "grad_norm": 0.08261553198099136, + "learning_rate": 0.00016448805141324016, + "loss": 0.2775, + "step": 34150 + }, + { + "epoch": 0.5328508142509515, + "grad_norm": 0.1017974391579628, + "learning_rate": 0.00016447765229508538, + "loss": 0.2095, + "step": 34160 + }, + { + "epoch": 0.5330068010232732, + "grad_norm": 1.3759571313858032, + "learning_rate": 0.0001644672531769306, + "loss": 0.2643, + "step": 34170 + }, + { + "epoch": 0.5331627877955949, + "grad_norm": 1.2654389142990112, + "learning_rate": 0.00016445685405877582, + "loss": 0.2949, + "step": 34180 + }, + { + "epoch": 0.5333187745679167, + "grad_norm": 1.5481843948364258, + "learning_rate": 0.00016444645494062104, + "loss": 0.1264, + "step": 34190 + }, + { + "epoch": 0.5334747613402383, + "grad_norm": 1.8094528913497925, + "learning_rate": 0.00016443605582246626, + "loss": 0.2727, + "step": 34200 + }, + { + "epoch": 0.5336307481125601, + "grad_norm": 2.224538564682007, + "learning_rate": 0.00016442565670431148, + "loss": 0.3096, + "step": 34210 + }, + { + "epoch": 0.5337867348848817, + "grad_norm": 0.6375226974487305, + "learning_rate": 0.0001644152575861567, + "loss": 0.2251, + "step": 34220 + }, + { + "epoch": 0.5339427216572035, + "grad_norm": 3.727106809616089, + "learning_rate": 0.00016440485846800192, + "loss": 0.4374, + "step": 34230 + }, + { + "epoch": 0.5340987084295252, + "grad_norm": 0.13345426321029663, + "learning_rate": 0.00016439445934984714, + "loss": 0.2011, + "step": 34240 + }, + { + "epoch": 0.5342546952018469, + "grad_norm": 2.1658668518066406, + "learning_rate": 0.00016438406023169236, + "loss": 0.2457, + "step": 34250 + }, + { + "epoch": 0.5344106819741686, + "grad_norm": 19.238407135009766, + "learning_rate": 0.00016437366111353758, + "loss": 0.2756, + "step": 34260 + }, + { + "epoch": 0.5345666687464903, + "grad_norm": 1.0292778015136719, + "learning_rate": 0.0001643632619953828, + "loss": 0.1646, + "step": 34270 + }, + { + "epoch": 0.534722655518812, + "grad_norm": 0.9372987747192383, + "learning_rate": 0.00016435286287722802, + "loss": 0.2762, + "step": 34280 + }, + { + "epoch": 0.5348786422911337, + "grad_norm": 0.3918002247810364, + "learning_rate": 0.00016434246375907324, + "loss": 0.1406, + "step": 34290 + }, + { + "epoch": 0.5350346290634554, + "grad_norm": 1.3518732786178589, + "learning_rate": 0.00016433206464091846, + "loss": 0.2992, + "step": 34300 + }, + { + "epoch": 0.5351906158357771, + "grad_norm": 0.73117595911026, + "learning_rate": 0.00016432166552276368, + "loss": 0.1172, + "step": 34310 + }, + { + "epoch": 0.5353466026080989, + "grad_norm": 0.06655958294868469, + "learning_rate": 0.0001643112664046089, + "loss": 0.1332, + "step": 34320 + }, + { + "epoch": 0.5355025893804205, + "grad_norm": 3.2587168216705322, + "learning_rate": 0.00016430086728645412, + "loss": 0.2052, + "step": 34330 + }, + { + "epoch": 0.5356585761527423, + "grad_norm": 2.6668431758880615, + "learning_rate": 0.00016429046816829934, + "loss": 0.1295, + "step": 34340 + }, + { + "epoch": 0.5358145629250639, + "grad_norm": 1.316080927848816, + "learning_rate": 0.00016428006905014456, + "loss": 0.2072, + "step": 34350 + }, + { + "epoch": 0.5359705496973857, + "grad_norm": 5.144528388977051, + "learning_rate": 0.00016426966993198977, + "loss": 0.1934, + "step": 34360 + }, + { + "epoch": 0.5361265364697073, + "grad_norm": 1.3529599905014038, + "learning_rate": 0.000164259270813835, + "loss": 0.2077, + "step": 34370 + }, + { + "epoch": 0.5362825232420291, + "grad_norm": 0.9132925868034363, + "learning_rate": 0.00016424887169568021, + "loss": 0.2404, + "step": 34380 + }, + { + "epoch": 0.5364385100143508, + "grad_norm": 1.520033597946167, + "learning_rate": 0.00016423847257752543, + "loss": 0.1765, + "step": 34390 + }, + { + "epoch": 0.5365944967866725, + "grad_norm": 2.343975305557251, + "learning_rate": 0.00016422807345937065, + "loss": 0.1959, + "step": 34400 + }, + { + "epoch": 0.5367504835589942, + "grad_norm": 0.5855118632316589, + "learning_rate": 0.00016421767434121587, + "loss": 0.0848, + "step": 34410 + }, + { + "epoch": 0.5369064703313159, + "grad_norm": 1.9108648300170898, + "learning_rate": 0.0001642072752230611, + "loss": 0.2406, + "step": 34420 + }, + { + "epoch": 0.5370624571036376, + "grad_norm": 1.0323792695999146, + "learning_rate": 0.0001641968761049063, + "loss": 0.2298, + "step": 34430 + }, + { + "epoch": 0.5372184438759593, + "grad_norm": 0.17435620725154877, + "learning_rate": 0.00016418647698675153, + "loss": 0.1396, + "step": 34440 + }, + { + "epoch": 0.537374430648281, + "grad_norm": 0.3526577353477478, + "learning_rate": 0.00016417607786859675, + "loss": 0.172, + "step": 34450 + }, + { + "epoch": 0.5375304174206027, + "grad_norm": 2.73685884475708, + "learning_rate": 0.00016416567875044197, + "loss": 0.1167, + "step": 34460 + }, + { + "epoch": 0.5376864041929245, + "grad_norm": 2.765693426132202, + "learning_rate": 0.0001641552796322872, + "loss": 0.1356, + "step": 34470 + }, + { + "epoch": 0.5378423909652461, + "grad_norm": 0.47628021240234375, + "learning_rate": 0.0001641448805141324, + "loss": 0.3104, + "step": 34480 + }, + { + "epoch": 0.5379983777375679, + "grad_norm": 1.1307590007781982, + "learning_rate": 0.00016413448139597763, + "loss": 0.3247, + "step": 34490 + }, + { + "epoch": 0.5381543645098895, + "grad_norm": 0.07136381417512894, + "learning_rate": 0.00016412408227782285, + "loss": 0.319, + "step": 34500 + }, + { + "epoch": 0.5383103512822113, + "grad_norm": 1.7850221395492554, + "learning_rate": 0.00016411368315966807, + "loss": 0.2515, + "step": 34510 + }, + { + "epoch": 0.5384663380545329, + "grad_norm": 1.3870742321014404, + "learning_rate": 0.0001641032840415133, + "loss": 0.256, + "step": 34520 + }, + { + "epoch": 0.5386223248268547, + "grad_norm": 1.1329221725463867, + "learning_rate": 0.0001640928849233585, + "loss": 0.2255, + "step": 34530 + }, + { + "epoch": 0.5387783115991764, + "grad_norm": 0.23236137628555298, + "learning_rate": 0.00016408248580520373, + "loss": 0.4053, + "step": 34540 + }, + { + "epoch": 0.5389342983714981, + "grad_norm": 3.350924253463745, + "learning_rate": 0.00016407208668704895, + "loss": 0.2816, + "step": 34550 + }, + { + "epoch": 0.5390902851438198, + "grad_norm": 1.7459170818328857, + "learning_rate": 0.00016406168756889417, + "loss": 0.0834, + "step": 34560 + }, + { + "epoch": 0.5392462719161415, + "grad_norm": 0.89351487159729, + "learning_rate": 0.0001640512884507394, + "loss": 0.1146, + "step": 34570 + }, + { + "epoch": 0.5394022586884633, + "grad_norm": 0.49410831928253174, + "learning_rate": 0.0001640408893325846, + "loss": 0.2319, + "step": 34580 + }, + { + "epoch": 0.5395582454607849, + "grad_norm": 0.009971237741410732, + "learning_rate": 0.00016403049021442983, + "loss": 0.1674, + "step": 34590 + }, + { + "epoch": 0.5397142322331067, + "grad_norm": 1.420511245727539, + "learning_rate": 0.00016402009109627505, + "loss": 0.1469, + "step": 34600 + }, + { + "epoch": 0.5398702190054283, + "grad_norm": 2.5686769485473633, + "learning_rate": 0.00016400969197812027, + "loss": 0.2607, + "step": 34610 + }, + { + "epoch": 0.5400262057777501, + "grad_norm": 2.5808913707733154, + "learning_rate": 0.00016399929285996549, + "loss": 0.4412, + "step": 34620 + }, + { + "epoch": 0.5401821925500717, + "grad_norm": 0.2778591811656952, + "learning_rate": 0.0001639888937418107, + "loss": 0.1563, + "step": 34630 + }, + { + "epoch": 0.5403381793223935, + "grad_norm": 1.6448099613189697, + "learning_rate": 0.00016397849462365592, + "loss": 0.2484, + "step": 34640 + }, + { + "epoch": 0.5404941660947151, + "grad_norm": 0.003213417250663042, + "learning_rate": 0.00016396809550550114, + "loss": 0.1582, + "step": 34650 + }, + { + "epoch": 0.5406501528670369, + "grad_norm": 2.3433990478515625, + "learning_rate": 0.00016395769638734636, + "loss": 0.2774, + "step": 34660 + }, + { + "epoch": 0.5408061396393585, + "grad_norm": 0.09544426947832108, + "learning_rate": 0.00016394729726919158, + "loss": 0.1242, + "step": 34670 + }, + { + "epoch": 0.5409621264116803, + "grad_norm": 1.8934235572814941, + "learning_rate": 0.0001639368981510368, + "loss": 0.1636, + "step": 34680 + }, + { + "epoch": 0.541118113184002, + "grad_norm": 2.8528554439544678, + "learning_rate": 0.00016392649903288202, + "loss": 0.2524, + "step": 34690 + }, + { + "epoch": 0.5412740999563237, + "grad_norm": 2.837836980819702, + "learning_rate": 0.00016391609991472722, + "loss": 0.1808, + "step": 34700 + }, + { + "epoch": 0.5414300867286455, + "grad_norm": 4.349740028381348, + "learning_rate": 0.00016390570079657246, + "loss": 0.12, + "step": 34710 + }, + { + "epoch": 0.5415860735009671, + "grad_norm": 0.6011945605278015, + "learning_rate": 0.00016389530167841765, + "loss": 0.2865, + "step": 34720 + }, + { + "epoch": 0.5417420602732889, + "grad_norm": 1.2079488039016724, + "learning_rate": 0.0001638849025602629, + "loss": 0.125, + "step": 34730 + }, + { + "epoch": 0.5418980470456105, + "grad_norm": 1.9549164772033691, + "learning_rate": 0.0001638745034421081, + "loss": 0.3285, + "step": 34740 + }, + { + "epoch": 0.5420540338179323, + "grad_norm": 0.6969407200813293, + "learning_rate": 0.00016386410432395334, + "loss": 0.3267, + "step": 34750 + }, + { + "epoch": 0.5422100205902539, + "grad_norm": 1.0561965703964233, + "learning_rate": 0.00016385370520579853, + "loss": 0.1463, + "step": 34760 + }, + { + "epoch": 0.5423660073625757, + "grad_norm": 1.8208931684494019, + "learning_rate": 0.00016384330608764378, + "loss": 0.1713, + "step": 34770 + }, + { + "epoch": 0.5425219941348973, + "grad_norm": 0.6349910497665405, + "learning_rate": 0.000163832906969489, + "loss": 0.2244, + "step": 34780 + }, + { + "epoch": 0.5426779809072191, + "grad_norm": 1.1976515054702759, + "learning_rate": 0.00016382250785133422, + "loss": 0.3407, + "step": 34790 + }, + { + "epoch": 0.5428339676795407, + "grad_norm": 0.5970319509506226, + "learning_rate": 0.00016381210873317944, + "loss": 0.1419, + "step": 34800 + }, + { + "epoch": 0.5429899544518625, + "grad_norm": 0.14623159170150757, + "learning_rate": 0.00016380170961502466, + "loss": 0.14, + "step": 34810 + }, + { + "epoch": 0.5431459412241841, + "grad_norm": 1.4332351684570312, + "learning_rate": 0.00016379131049686988, + "loss": 0.1093, + "step": 34820 + }, + { + "epoch": 0.5433019279965059, + "grad_norm": 1.1443063020706177, + "learning_rate": 0.0001637809113787151, + "loss": 0.3676, + "step": 34830 + }, + { + "epoch": 0.5434579147688277, + "grad_norm": 1.3897461891174316, + "learning_rate": 0.00016377051226056032, + "loss": 0.3713, + "step": 34840 + }, + { + "epoch": 0.5436139015411493, + "grad_norm": 1.7773199081420898, + "learning_rate": 0.00016376011314240554, + "loss": 0.2249, + "step": 34850 + }, + { + "epoch": 0.5437698883134711, + "grad_norm": 0.8524389266967773, + "learning_rate": 0.00016374971402425076, + "loss": 0.2158, + "step": 34860 + }, + { + "epoch": 0.5439258750857927, + "grad_norm": 0.5572389960289001, + "learning_rate": 0.00016373931490609598, + "loss": 0.215, + "step": 34870 + }, + { + "epoch": 0.5440818618581145, + "grad_norm": 3.0238709449768066, + "learning_rate": 0.0001637289157879412, + "loss": 0.3765, + "step": 34880 + }, + { + "epoch": 0.5442378486304361, + "grad_norm": 1.1655421257019043, + "learning_rate": 0.00016371851666978642, + "loss": 0.2491, + "step": 34890 + }, + { + "epoch": 0.5443938354027579, + "grad_norm": 0.49492090940475464, + "learning_rate": 0.00016370811755163164, + "loss": 0.1287, + "step": 34900 + }, + { + "epoch": 0.5445498221750795, + "grad_norm": 0.2732921540737152, + "learning_rate": 0.00016369771843347685, + "loss": 0.2154, + "step": 34910 + }, + { + "epoch": 0.5447058089474013, + "grad_norm": 2.5807769298553467, + "learning_rate": 0.00016368731931532207, + "loss": 0.2032, + "step": 34920 + }, + { + "epoch": 0.5448617957197229, + "grad_norm": 0.5044315457344055, + "learning_rate": 0.0001636769201971673, + "loss": 0.2595, + "step": 34930 + }, + { + "epoch": 0.5450177824920447, + "grad_norm": 2.1248972415924072, + "learning_rate": 0.00016366652107901251, + "loss": 0.5651, + "step": 34940 + }, + { + "epoch": 0.5451737692643663, + "grad_norm": 0.8391468524932861, + "learning_rate": 0.00016365612196085773, + "loss": 0.3028, + "step": 34950 + }, + { + "epoch": 0.5453297560366881, + "grad_norm": 0.36081477999687195, + "learning_rate": 0.00016364572284270295, + "loss": 0.1073, + "step": 34960 + }, + { + "epoch": 0.5454857428090097, + "grad_norm": 3.175804376602173, + "learning_rate": 0.00016363532372454817, + "loss": 0.1889, + "step": 34970 + }, + { + "epoch": 0.5456417295813315, + "grad_norm": 0.20647937059402466, + "learning_rate": 0.0001636249246063934, + "loss": 0.2416, + "step": 34980 + }, + { + "epoch": 0.5457977163536533, + "grad_norm": 2.3525943756103516, + "learning_rate": 0.0001636145254882386, + "loss": 0.3179, + "step": 34990 + }, + { + "epoch": 0.5459537031259749, + "grad_norm": 0.2641216814517975, + "learning_rate": 0.00016360412637008383, + "loss": 0.2157, + "step": 35000 + }, + { + "epoch": 0.5461096898982967, + "grad_norm": 0.09906073659658432, + "learning_rate": 0.00016359372725192905, + "loss": 0.2576, + "step": 35010 + }, + { + "epoch": 0.5462656766706183, + "grad_norm": 0.13971920311450958, + "learning_rate": 0.00016358332813377427, + "loss": 0.1298, + "step": 35020 + }, + { + "epoch": 0.5464216634429401, + "grad_norm": 1.4535638093948364, + "learning_rate": 0.0001635729290156195, + "loss": 0.2357, + "step": 35030 + }, + { + "epoch": 0.5465776502152617, + "grad_norm": 0.9107828736305237, + "learning_rate": 0.0001635625298974647, + "loss": 0.3049, + "step": 35040 + }, + { + "epoch": 0.5467336369875835, + "grad_norm": 0.0377386212348938, + "learning_rate": 0.00016355213077930993, + "loss": 0.2066, + "step": 35050 + }, + { + "epoch": 0.5468896237599051, + "grad_norm": 2.7800869941711426, + "learning_rate": 0.00016354173166115515, + "loss": 0.2974, + "step": 35060 + }, + { + "epoch": 0.5470456105322269, + "grad_norm": 1.6247998476028442, + "learning_rate": 0.00016353133254300037, + "loss": 0.3822, + "step": 35070 + }, + { + "epoch": 0.5472015973045485, + "grad_norm": 1.6479015350341797, + "learning_rate": 0.0001635209334248456, + "loss": 0.3401, + "step": 35080 + }, + { + "epoch": 0.5473575840768703, + "grad_norm": 0.1879737675189972, + "learning_rate": 0.0001635105343066908, + "loss": 0.1549, + "step": 35090 + }, + { + "epoch": 0.5475135708491919, + "grad_norm": 0.39355361461639404, + "learning_rate": 0.00016350013518853603, + "loss": 0.2032, + "step": 35100 + }, + { + "epoch": 0.5476695576215137, + "grad_norm": 2.9976983070373535, + "learning_rate": 0.00016348973607038125, + "loss": 0.3867, + "step": 35110 + }, + { + "epoch": 0.5478255443938354, + "grad_norm": 0.8242707252502441, + "learning_rate": 0.00016347933695222647, + "loss": 0.3601, + "step": 35120 + }, + { + "epoch": 0.5479815311661571, + "grad_norm": 0.874955415725708, + "learning_rate": 0.0001634689378340717, + "loss": 0.2709, + "step": 35130 + }, + { + "epoch": 0.5481375179384789, + "grad_norm": 0.05980971083045006, + "learning_rate": 0.0001634585387159169, + "loss": 0.2646, + "step": 35140 + }, + { + "epoch": 0.5482935047108005, + "grad_norm": 1.8437328338623047, + "learning_rate": 0.0001634481395977621, + "loss": 0.1965, + "step": 35150 + }, + { + "epoch": 0.5484494914831223, + "grad_norm": 0.4142405688762665, + "learning_rate": 0.00016343774047960735, + "loss": 0.1756, + "step": 35160 + }, + { + "epoch": 0.5486054782554439, + "grad_norm": 4.836324214935303, + "learning_rate": 0.00016342734136145254, + "loss": 0.2534, + "step": 35170 + }, + { + "epoch": 0.5487614650277657, + "grad_norm": 1.4933065176010132, + "learning_rate": 0.00016341694224329779, + "loss": 0.2355, + "step": 35180 + }, + { + "epoch": 0.5489174518000873, + "grad_norm": 1.9468894004821777, + "learning_rate": 0.00016340654312514298, + "loss": 0.3113, + "step": 35190 + }, + { + "epoch": 0.5490734385724091, + "grad_norm": 1.012710452079773, + "learning_rate": 0.00016339614400698822, + "loss": 0.2635, + "step": 35200 + }, + { + "epoch": 0.5492294253447307, + "grad_norm": 1.5551140308380127, + "learning_rate": 0.00016338574488883342, + "loss": 0.2951, + "step": 35210 + }, + { + "epoch": 0.5493854121170525, + "grad_norm": 1.8613696098327637, + "learning_rate": 0.00016337534577067866, + "loss": 0.2832, + "step": 35220 + }, + { + "epoch": 0.5495413988893741, + "grad_norm": 1.6748839616775513, + "learning_rate": 0.00016336494665252386, + "loss": 0.2266, + "step": 35230 + }, + { + "epoch": 0.5496973856616959, + "grad_norm": 0.4094032347202301, + "learning_rate": 0.0001633545475343691, + "loss": 0.3497, + "step": 35240 + }, + { + "epoch": 0.5498533724340176, + "grad_norm": 1.6414631605148315, + "learning_rate": 0.0001633441484162143, + "loss": 0.1763, + "step": 35250 + }, + { + "epoch": 0.5500093592063393, + "grad_norm": 1.6380645036697388, + "learning_rate": 0.00016333374929805954, + "loss": 0.1355, + "step": 35260 + }, + { + "epoch": 0.550165345978661, + "grad_norm": 0.7325630187988281, + "learning_rate": 0.00016332335017990473, + "loss": 0.3105, + "step": 35270 + }, + { + "epoch": 0.5503213327509827, + "grad_norm": 1.6015644073486328, + "learning_rate": 0.00016331295106174998, + "loss": 0.2311, + "step": 35280 + }, + { + "epoch": 0.5504773195233045, + "grad_norm": 0.9224210381507874, + "learning_rate": 0.00016330255194359517, + "loss": 0.2746, + "step": 35290 + }, + { + "epoch": 0.5506333062956261, + "grad_norm": 1.30025315284729, + "learning_rate": 0.00016329215282544042, + "loss": 0.2066, + "step": 35300 + }, + { + "epoch": 0.5507892930679479, + "grad_norm": 0.481125146150589, + "learning_rate": 0.0001632817537072856, + "loss": 0.1778, + "step": 35310 + }, + { + "epoch": 0.5509452798402695, + "grad_norm": 0.2709486782550812, + "learning_rate": 0.00016327135458913086, + "loss": 0.1131, + "step": 35320 + }, + { + "epoch": 0.5511012666125913, + "grad_norm": 0.6248563528060913, + "learning_rate": 0.00016326095547097605, + "loss": 0.4356, + "step": 35330 + }, + { + "epoch": 0.5512572533849129, + "grad_norm": 3.0947647094726562, + "learning_rate": 0.0001632505563528213, + "loss": 0.3072, + "step": 35340 + }, + { + "epoch": 0.5514132401572347, + "grad_norm": 1.009535312652588, + "learning_rate": 0.0001632401572346665, + "loss": 0.1808, + "step": 35350 + }, + { + "epoch": 0.5515692269295563, + "grad_norm": 0.042604975402355194, + "learning_rate": 0.00016322975811651174, + "loss": 0.1198, + "step": 35360 + }, + { + "epoch": 0.5517252137018781, + "grad_norm": 1.8652396202087402, + "learning_rate": 0.00016321935899835693, + "loss": 0.2176, + "step": 35370 + }, + { + "epoch": 0.5518812004741998, + "grad_norm": 4.6097187995910645, + "learning_rate": 0.00016320895988020218, + "loss": 0.2132, + "step": 35380 + }, + { + "epoch": 0.5520371872465215, + "grad_norm": 1.3615522384643555, + "learning_rate": 0.00016319856076204737, + "loss": 0.2751, + "step": 35390 + }, + { + "epoch": 0.5521931740188432, + "grad_norm": 0.8844773769378662, + "learning_rate": 0.00016318816164389262, + "loss": 0.22, + "step": 35400 + }, + { + "epoch": 0.5523491607911649, + "grad_norm": 0.5413331985473633, + "learning_rate": 0.0001631777625257378, + "loss": 0.2159, + "step": 35410 + }, + { + "epoch": 0.5525051475634866, + "grad_norm": 1.5456678867340088, + "learning_rate": 0.00016316736340758306, + "loss": 0.2244, + "step": 35420 + }, + { + "epoch": 0.5526611343358083, + "grad_norm": 2.1405861377716064, + "learning_rate": 0.00016315696428942825, + "loss": 0.446, + "step": 35430 + }, + { + "epoch": 0.5528171211081301, + "grad_norm": 1.4269858598709106, + "learning_rate": 0.0001631465651712735, + "loss": 0.3599, + "step": 35440 + }, + { + "epoch": 0.5529731078804517, + "grad_norm": 1.8682516813278198, + "learning_rate": 0.0001631361660531187, + "loss": 0.115, + "step": 35450 + }, + { + "epoch": 0.5531290946527735, + "grad_norm": 1.0175774097442627, + "learning_rate": 0.00016312576693496394, + "loss": 0.4003, + "step": 35460 + }, + { + "epoch": 0.5532850814250951, + "grad_norm": 2.4369170665740967, + "learning_rate": 0.00016311536781680913, + "loss": 0.2489, + "step": 35470 + }, + { + "epoch": 0.5534410681974169, + "grad_norm": 1.886022686958313, + "learning_rate": 0.00016310496869865437, + "loss": 0.4201, + "step": 35480 + }, + { + "epoch": 0.5535970549697385, + "grad_norm": 1.5811959505081177, + "learning_rate": 0.00016309456958049957, + "loss": 0.4286, + "step": 35490 + }, + { + "epoch": 0.5537530417420603, + "grad_norm": 1.971110224723816, + "learning_rate": 0.00016308417046234481, + "loss": 0.3599, + "step": 35500 + }, + { + "epoch": 0.553909028514382, + "grad_norm": 0.0335380844771862, + "learning_rate": 0.00016307377134419, + "loss": 0.3357, + "step": 35510 + }, + { + "epoch": 0.5540650152867037, + "grad_norm": 1.8519577980041504, + "learning_rate": 0.00016306337222603525, + "loss": 0.3126, + "step": 35520 + }, + { + "epoch": 0.5542210020590254, + "grad_norm": 2.1563379764556885, + "learning_rate": 0.00016305297310788045, + "loss": 0.1123, + "step": 35530 + }, + { + "epoch": 0.5543769888313471, + "grad_norm": 1.64332914352417, + "learning_rate": 0.0001630425739897257, + "loss": 0.2853, + "step": 35540 + }, + { + "epoch": 0.5545329756036688, + "grad_norm": 0.061150554567575455, + "learning_rate": 0.00016303217487157088, + "loss": 0.1943, + "step": 35550 + }, + { + "epoch": 0.5546889623759905, + "grad_norm": 1.2701060771942139, + "learning_rate": 0.00016302177575341613, + "loss": 0.2304, + "step": 35560 + }, + { + "epoch": 0.5548449491483122, + "grad_norm": 2.424860715866089, + "learning_rate": 0.00016301137663526132, + "loss": 0.2128, + "step": 35570 + }, + { + "epoch": 0.5550009359206339, + "grad_norm": 0.6803575158119202, + "learning_rate": 0.00016300097751710657, + "loss": 0.1145, + "step": 35580 + }, + { + "epoch": 0.5551569226929557, + "grad_norm": 1.2855092287063599, + "learning_rate": 0.00016299057839895176, + "loss": 0.2448, + "step": 35590 + }, + { + "epoch": 0.5553129094652773, + "grad_norm": 2.6340911388397217, + "learning_rate": 0.00016298017928079698, + "loss": 0.1856, + "step": 35600 + }, + { + "epoch": 0.5554688962375991, + "grad_norm": 1.776382327079773, + "learning_rate": 0.0001629697801626422, + "loss": 0.2422, + "step": 35610 + }, + { + "epoch": 0.5556248830099207, + "grad_norm": 3.0746347904205322, + "learning_rate": 0.00016295938104448742, + "loss": 0.1578, + "step": 35620 + }, + { + "epoch": 0.5557808697822425, + "grad_norm": 1.4887659549713135, + "learning_rate": 0.00016294898192633267, + "loss": 0.2385, + "step": 35630 + }, + { + "epoch": 0.5559368565545642, + "grad_norm": 0.5371220707893372, + "learning_rate": 0.00016293858280817786, + "loss": 0.2361, + "step": 35640 + }, + { + "epoch": 0.5560928433268859, + "grad_norm": 0.18604904413223267, + "learning_rate": 0.0001629281836900231, + "loss": 0.0828, + "step": 35650 + }, + { + "epoch": 0.5562488300992076, + "grad_norm": 2.267854690551758, + "learning_rate": 0.0001629177845718683, + "loss": 0.413, + "step": 35660 + }, + { + "epoch": 0.5564048168715293, + "grad_norm": 1.9755452871322632, + "learning_rate": 0.00016290738545371355, + "loss": 0.146, + "step": 35670 + }, + { + "epoch": 0.556560803643851, + "grad_norm": 2.430293321609497, + "learning_rate": 0.00016289698633555874, + "loss": 0.2677, + "step": 35680 + }, + { + "epoch": 0.5567167904161727, + "grad_norm": 0.6319543719291687, + "learning_rate": 0.000162886587217404, + "loss": 0.1481, + "step": 35690 + }, + { + "epoch": 0.5568727771884944, + "grad_norm": 0.17439277470111847, + "learning_rate": 0.00016287618809924918, + "loss": 0.1641, + "step": 35700 + }, + { + "epoch": 0.5570287639608161, + "grad_norm": 0.350175142288208, + "learning_rate": 0.00016286578898109443, + "loss": 0.393, + "step": 35710 + }, + { + "epoch": 0.5571847507331378, + "grad_norm": 2.641941547393799, + "learning_rate": 0.00016285538986293962, + "loss": 0.1907, + "step": 35720 + }, + { + "epoch": 0.5573407375054595, + "grad_norm": 0.1910303384065628, + "learning_rate": 0.00016284499074478487, + "loss": 0.2248, + "step": 35730 + }, + { + "epoch": 0.5574967242777813, + "grad_norm": 1.0180896520614624, + "learning_rate": 0.00016283459162663006, + "loss": 0.1924, + "step": 35740 + }, + { + "epoch": 0.557652711050103, + "grad_norm": 1.2047260999679565, + "learning_rate": 0.0001628241925084753, + "loss": 0.2031, + "step": 35750 + }, + { + "epoch": 0.5578086978224247, + "grad_norm": 1.8182405233383179, + "learning_rate": 0.0001628137933903205, + "loss": 0.1573, + "step": 35760 + }, + { + "epoch": 0.5579646845947464, + "grad_norm": 0.5485963821411133, + "learning_rate": 0.00016280339427216574, + "loss": 0.152, + "step": 35770 + }, + { + "epoch": 0.5581206713670681, + "grad_norm": 0.1674145758152008, + "learning_rate": 0.00016279299515401094, + "loss": 0.2388, + "step": 35780 + }, + { + "epoch": 0.5582766581393898, + "grad_norm": 0.36187657713890076, + "learning_rate": 0.00016278259603585618, + "loss": 0.1618, + "step": 35790 + }, + { + "epoch": 0.5584326449117115, + "grad_norm": 0.013768521137535572, + "learning_rate": 0.00016277219691770138, + "loss": 0.3666, + "step": 35800 + }, + { + "epoch": 0.5585886316840332, + "grad_norm": 2.8602802753448486, + "learning_rate": 0.00016276179779954662, + "loss": 0.2065, + "step": 35810 + }, + { + "epoch": 0.5587446184563549, + "grad_norm": 1.7925455570220947, + "learning_rate": 0.00016275139868139182, + "loss": 0.1729, + "step": 35820 + }, + { + "epoch": 0.5589006052286766, + "grad_norm": 2.910456418991089, + "learning_rate": 0.00016274099956323706, + "loss": 0.3734, + "step": 35830 + }, + { + "epoch": 0.5590565920009983, + "grad_norm": 0.8378308415412903, + "learning_rate": 0.00016273060044508225, + "loss": 0.2255, + "step": 35840 + }, + { + "epoch": 0.55921257877332, + "grad_norm": 0.409534752368927, + "learning_rate": 0.0001627202013269275, + "loss": 0.2046, + "step": 35850 + }, + { + "epoch": 0.5593685655456417, + "grad_norm": 0.49498099088668823, + "learning_rate": 0.0001627098022087727, + "loss": 0.1058, + "step": 35860 + }, + { + "epoch": 0.5595245523179634, + "grad_norm": 1.1761783361434937, + "learning_rate": 0.00016269940309061794, + "loss": 0.1566, + "step": 35870 + }, + { + "epoch": 0.5596805390902851, + "grad_norm": 3.8204751014709473, + "learning_rate": 0.00016268900397246313, + "loss": 0.2646, + "step": 35880 + }, + { + "epoch": 0.5598365258626069, + "grad_norm": 0.9882522225379944, + "learning_rate": 0.00016267860485430838, + "loss": 0.1756, + "step": 35890 + }, + { + "epoch": 0.5599925126349286, + "grad_norm": 1.1832259893417358, + "learning_rate": 0.00016266820573615357, + "loss": 0.1385, + "step": 35900 + }, + { + "epoch": 0.5601484994072503, + "grad_norm": 0.7638296484947205, + "learning_rate": 0.00016265780661799882, + "loss": 0.5679, + "step": 35910 + }, + { + "epoch": 0.560304486179572, + "grad_norm": 0.8551504611968994, + "learning_rate": 0.000162647407499844, + "loss": 0.2981, + "step": 35920 + }, + { + "epoch": 0.5604604729518937, + "grad_norm": 5.499948501586914, + "learning_rate": 0.00016263700838168926, + "loss": 0.1429, + "step": 35930 + }, + { + "epoch": 0.5606164597242154, + "grad_norm": 3.0494496822357178, + "learning_rate": 0.00016262660926353445, + "loss": 0.1322, + "step": 35940 + }, + { + "epoch": 0.5607724464965371, + "grad_norm": 0.2482384592294693, + "learning_rate": 0.0001626162101453797, + "loss": 0.2473, + "step": 35950 + }, + { + "epoch": 0.5609284332688588, + "grad_norm": 0.009750776924192905, + "learning_rate": 0.0001626058110272249, + "loss": 0.0429, + "step": 35960 + }, + { + "epoch": 0.5610844200411805, + "grad_norm": 0.5742604732513428, + "learning_rate": 0.00016259541190907014, + "loss": 0.4565, + "step": 35970 + }, + { + "epoch": 0.5612404068135022, + "grad_norm": 0.7810243368148804, + "learning_rate": 0.00016258501279091533, + "loss": 0.2107, + "step": 35980 + }, + { + "epoch": 0.5613963935858239, + "grad_norm": 2.785747528076172, + "learning_rate": 0.00016257461367276058, + "loss": 0.5087, + "step": 35990 + }, + { + "epoch": 0.5615523803581456, + "grad_norm": 0.8806902766227722, + "learning_rate": 0.00016256421455460577, + "loss": 0.2937, + "step": 36000 + }, + { + "epoch": 0.5617083671304673, + "grad_norm": 1.857373833656311, + "learning_rate": 0.00016255381543645102, + "loss": 0.5042, + "step": 36010 + }, + { + "epoch": 0.561864353902789, + "grad_norm": 0.5333901047706604, + "learning_rate": 0.0001625434163182962, + "loss": 0.1678, + "step": 36020 + }, + { + "epoch": 0.5620203406751108, + "grad_norm": 11.023160934448242, + "learning_rate": 0.00016253301720014145, + "loss": 0.2721, + "step": 36030 + }, + { + "epoch": 0.5621763274474325, + "grad_norm": 2.6214029788970947, + "learning_rate": 0.00016252261808198665, + "loss": 0.1675, + "step": 36040 + }, + { + "epoch": 0.5623323142197542, + "grad_norm": 0.25980064272880554, + "learning_rate": 0.0001625122189638319, + "loss": 0.1832, + "step": 36050 + }, + { + "epoch": 0.5624883009920759, + "grad_norm": 1.3559473752975464, + "learning_rate": 0.0001625018198456771, + "loss": 0.121, + "step": 36060 + }, + { + "epoch": 0.5626442877643976, + "grad_norm": 0.3558153212070465, + "learning_rate": 0.0001624914207275223, + "loss": 0.2797, + "step": 36070 + }, + { + "epoch": 0.5628002745367193, + "grad_norm": 0.6314427256584167, + "learning_rate": 0.00016248102160936753, + "loss": 0.3455, + "step": 36080 + }, + { + "epoch": 0.562956261309041, + "grad_norm": 0.16120101511478424, + "learning_rate": 0.00016247062249121275, + "loss": 0.3669, + "step": 36090 + }, + { + "epoch": 0.5631122480813627, + "grad_norm": 1.041572093963623, + "learning_rate": 0.00016246022337305797, + "loss": 0.2744, + "step": 36100 + }, + { + "epoch": 0.5632682348536844, + "grad_norm": 0.2645890414714813, + "learning_rate": 0.00016244982425490318, + "loss": 0.1487, + "step": 36110 + }, + { + "epoch": 0.5634242216260061, + "grad_norm": 5.698698043823242, + "learning_rate": 0.0001624394251367484, + "loss": 0.173, + "step": 36120 + }, + { + "epoch": 0.5635802083983278, + "grad_norm": 0.3299804627895355, + "learning_rate": 0.00016242902601859362, + "loss": 0.1965, + "step": 36130 + }, + { + "epoch": 0.5637361951706495, + "grad_norm": 1.2476481199264526, + "learning_rate": 0.00016241862690043884, + "loss": 0.4097, + "step": 36140 + }, + { + "epoch": 0.5638921819429712, + "grad_norm": 0.30926263332366943, + "learning_rate": 0.00016240822778228406, + "loss": 0.1416, + "step": 36150 + }, + { + "epoch": 0.564048168715293, + "grad_norm": 0.05872740224003792, + "learning_rate": 0.00016239782866412928, + "loss": 0.1449, + "step": 36160 + }, + { + "epoch": 0.5642041554876146, + "grad_norm": 0.2292211800813675, + "learning_rate": 0.0001623874295459745, + "loss": 0.2557, + "step": 36170 + }, + { + "epoch": 0.5643601422599364, + "grad_norm": 1.7822531461715698, + "learning_rate": 0.00016237703042781972, + "loss": 0.2953, + "step": 36180 + }, + { + "epoch": 0.5645161290322581, + "grad_norm": 2.4908461570739746, + "learning_rate": 0.00016236663130966494, + "loss": 0.3888, + "step": 36190 + }, + { + "epoch": 0.5646721158045798, + "grad_norm": 0.08363594114780426, + "learning_rate": 0.00016235623219151016, + "loss": 0.2293, + "step": 36200 + }, + { + "epoch": 0.5648281025769015, + "grad_norm": 1.4830002784729004, + "learning_rate": 0.00016234583307335538, + "loss": 0.2348, + "step": 36210 + }, + { + "epoch": 0.5649840893492232, + "grad_norm": 1.8443071842193604, + "learning_rate": 0.0001623354339552006, + "loss": 0.3392, + "step": 36220 + }, + { + "epoch": 0.5651400761215449, + "grad_norm": 2.8050875663757324, + "learning_rate": 0.00016232503483704582, + "loss": 0.2426, + "step": 36230 + }, + { + "epoch": 0.5652960628938666, + "grad_norm": 3.6627259254455566, + "learning_rate": 0.00016231463571889104, + "loss": 0.2118, + "step": 36240 + }, + { + "epoch": 0.5654520496661883, + "grad_norm": 0.3443094789981842, + "learning_rate": 0.00016230423660073626, + "loss": 0.2418, + "step": 36250 + }, + { + "epoch": 0.56560803643851, + "grad_norm": 1.7411112785339355, + "learning_rate": 0.00016229383748258148, + "loss": 0.1628, + "step": 36260 + }, + { + "epoch": 0.5657640232108317, + "grad_norm": 1.1321816444396973, + "learning_rate": 0.0001622834383644267, + "loss": 0.2732, + "step": 36270 + }, + { + "epoch": 0.5659200099831534, + "grad_norm": 2.9686615467071533, + "learning_rate": 0.00016227303924627192, + "loss": 0.3036, + "step": 36280 + }, + { + "epoch": 0.5660759967554752, + "grad_norm": 5.03530740737915, + "learning_rate": 0.00016226264012811714, + "loss": 0.2143, + "step": 36290 + }, + { + "epoch": 0.5662319835277968, + "grad_norm": 2.3943281173706055, + "learning_rate": 0.00016225224100996236, + "loss": 0.2629, + "step": 36300 + }, + { + "epoch": 0.5663879703001186, + "grad_norm": 1.812828779220581, + "learning_rate": 0.00016224184189180758, + "loss": 0.3892, + "step": 36310 + }, + { + "epoch": 0.5665439570724402, + "grad_norm": 1.4708483219146729, + "learning_rate": 0.0001622314427736528, + "loss": 0.2118, + "step": 36320 + }, + { + "epoch": 0.566699943844762, + "grad_norm": 8.99713134765625, + "learning_rate": 0.00016222104365549802, + "loss": 0.3185, + "step": 36330 + }, + { + "epoch": 0.5668559306170837, + "grad_norm": 1.7472341060638428, + "learning_rate": 0.00016221064453734324, + "loss": 0.1659, + "step": 36340 + }, + { + "epoch": 0.5670119173894054, + "grad_norm": 0.8059778213500977, + "learning_rate": 0.00016220024541918846, + "loss": 0.0621, + "step": 36350 + }, + { + "epoch": 0.5671679041617271, + "grad_norm": 4.0501861572265625, + "learning_rate": 0.00016218984630103368, + "loss": 0.1399, + "step": 36360 + }, + { + "epoch": 0.5673238909340488, + "grad_norm": 0.014400321058928967, + "learning_rate": 0.0001621794471828789, + "loss": 0.2372, + "step": 36370 + }, + { + "epoch": 0.5674798777063705, + "grad_norm": 0.4061933755874634, + "learning_rate": 0.00016216904806472412, + "loss": 0.3854, + "step": 36380 + }, + { + "epoch": 0.5676358644786922, + "grad_norm": 0.4940955340862274, + "learning_rate": 0.00016215864894656933, + "loss": 0.1689, + "step": 36390 + }, + { + "epoch": 0.567791851251014, + "grad_norm": 0.3220517635345459, + "learning_rate": 0.00016214824982841455, + "loss": 0.1912, + "step": 36400 + }, + { + "epoch": 0.5679478380233356, + "grad_norm": 2.0848920345306396, + "learning_rate": 0.00016213785071025977, + "loss": 0.1639, + "step": 36410 + }, + { + "epoch": 0.5681038247956574, + "grad_norm": 0.04129406809806824, + "learning_rate": 0.000162127451592105, + "loss": 0.1378, + "step": 36420 + }, + { + "epoch": 0.568259811567979, + "grad_norm": 1.9343401193618774, + "learning_rate": 0.0001621170524739502, + "loss": 0.3759, + "step": 36430 + }, + { + "epoch": 0.5684157983403008, + "grad_norm": 1.0625615119934082, + "learning_rate": 0.00016210665335579543, + "loss": 0.2074, + "step": 36440 + }, + { + "epoch": 0.5685717851126224, + "grad_norm": 0.14784491062164307, + "learning_rate": 0.00016209625423764065, + "loss": 0.0981, + "step": 36450 + }, + { + "epoch": 0.5687277718849442, + "grad_norm": 1.8117821216583252, + "learning_rate": 0.00016208585511948587, + "loss": 0.2755, + "step": 36460 + }, + { + "epoch": 0.5688837586572658, + "grad_norm": 0.061010006815195084, + "learning_rate": 0.0001620754560013311, + "loss": 0.1815, + "step": 36470 + }, + { + "epoch": 0.5690397454295876, + "grad_norm": 0.44453224539756775, + "learning_rate": 0.0001620650568831763, + "loss": 0.0891, + "step": 36480 + }, + { + "epoch": 0.5691957322019093, + "grad_norm": 0.6130645871162415, + "learning_rate": 0.00016205465776502153, + "loss": 0.2778, + "step": 36490 + }, + { + "epoch": 0.569351718974231, + "grad_norm": 5.408796787261963, + "learning_rate": 0.00016204425864686678, + "loss": 0.2301, + "step": 36500 + }, + { + "epoch": 0.5695077057465527, + "grad_norm": 1.2095412015914917, + "learning_rate": 0.00016203385952871197, + "loss": 0.1378, + "step": 36510 + }, + { + "epoch": 0.5696636925188744, + "grad_norm": 0.46873530745506287, + "learning_rate": 0.0001620234604105572, + "loss": 0.1431, + "step": 36520 + }, + { + "epoch": 0.5698196792911961, + "grad_norm": 1.028438925743103, + "learning_rate": 0.0001620130612924024, + "loss": 0.1114, + "step": 36530 + }, + { + "epoch": 0.5699756660635178, + "grad_norm": 0.5952587127685547, + "learning_rate": 0.00016200266217424763, + "loss": 0.2156, + "step": 36540 + }, + { + "epoch": 0.5701316528358396, + "grad_norm": 0.5134285688400269, + "learning_rate": 0.00016199226305609285, + "loss": 0.2712, + "step": 36550 + }, + { + "epoch": 0.5702876396081612, + "grad_norm": 0.03440163657069206, + "learning_rate": 0.00016198186393793807, + "loss": 0.312, + "step": 36560 + }, + { + "epoch": 0.570443626380483, + "grad_norm": 0.10363951325416565, + "learning_rate": 0.0001619714648197833, + "loss": 0.2399, + "step": 36570 + }, + { + "epoch": 0.5705996131528046, + "grad_norm": 1.3341947793960571, + "learning_rate": 0.0001619610657016285, + "loss": 0.1397, + "step": 36580 + }, + { + "epoch": 0.5707555999251264, + "grad_norm": 0.0668073296546936, + "learning_rate": 0.00016195066658347373, + "loss": 0.1091, + "step": 36590 + }, + { + "epoch": 0.570911586697448, + "grad_norm": 5.0486674308776855, + "learning_rate": 0.00016194026746531895, + "loss": 0.4893, + "step": 36600 + }, + { + "epoch": 0.5710675734697698, + "grad_norm": 1.2655692100524902, + "learning_rate": 0.00016192986834716417, + "loss": 0.3344, + "step": 36610 + }, + { + "epoch": 0.5712235602420914, + "grad_norm": 0.1458190381526947, + "learning_rate": 0.00016191946922900939, + "loss": 0.247, + "step": 36620 + }, + { + "epoch": 0.5713795470144132, + "grad_norm": 2.5789072513580322, + "learning_rate": 0.0001619090701108546, + "loss": 0.1546, + "step": 36630 + }, + { + "epoch": 0.5715355337867349, + "grad_norm": 1.191821575164795, + "learning_rate": 0.00016189867099269983, + "loss": 0.2215, + "step": 36640 + }, + { + "epoch": 0.5716915205590566, + "grad_norm": 8.163617134094238, + "learning_rate": 0.00016188827187454505, + "loss": 0.1696, + "step": 36650 + }, + { + "epoch": 0.5718475073313783, + "grad_norm": 3.162277936935425, + "learning_rate": 0.00016187787275639027, + "loss": 0.2906, + "step": 36660 + }, + { + "epoch": 0.5720034941037, + "grad_norm": 2.2368602752685547, + "learning_rate": 0.00016186747363823548, + "loss": 0.3806, + "step": 36670 + }, + { + "epoch": 0.5721594808760218, + "grad_norm": 2.0193803310394287, + "learning_rate": 0.0001618570745200807, + "loss": 0.1964, + "step": 36680 + }, + { + "epoch": 0.5723154676483434, + "grad_norm": 0.9131811261177063, + "learning_rate": 0.00016184667540192592, + "loss": 0.1902, + "step": 36690 + }, + { + "epoch": 0.5724714544206652, + "grad_norm": 0.21330870687961578, + "learning_rate": 0.00016183627628377114, + "loss": 0.2907, + "step": 36700 + }, + { + "epoch": 0.5726274411929868, + "grad_norm": 0.7145895957946777, + "learning_rate": 0.00016182587716561636, + "loss": 0.2574, + "step": 36710 + }, + { + "epoch": 0.5727834279653086, + "grad_norm": 1.0699433088302612, + "learning_rate": 0.00016181547804746158, + "loss": 0.1675, + "step": 36720 + }, + { + "epoch": 0.5729394147376302, + "grad_norm": 0.5066679120063782, + "learning_rate": 0.0001618050789293068, + "loss": 0.4487, + "step": 36730 + }, + { + "epoch": 0.573095401509952, + "grad_norm": 0.7482333183288574, + "learning_rate": 0.00016179467981115202, + "loss": 0.4101, + "step": 36740 + }, + { + "epoch": 0.5732513882822736, + "grad_norm": 0.15925532579421997, + "learning_rate": 0.00016178428069299724, + "loss": 0.2825, + "step": 36750 + }, + { + "epoch": 0.5734073750545954, + "grad_norm": 0.95689457654953, + "learning_rate": 0.00016177388157484246, + "loss": 0.2754, + "step": 36760 + }, + { + "epoch": 0.573563361826917, + "grad_norm": 2.15043568611145, + "learning_rate": 0.00016176348245668768, + "loss": 0.2073, + "step": 36770 + }, + { + "epoch": 0.5737193485992388, + "grad_norm": 2.651520252227783, + "learning_rate": 0.0001617530833385329, + "loss": 0.2517, + "step": 36780 + }, + { + "epoch": 0.5738753353715605, + "grad_norm": 2.1156702041625977, + "learning_rate": 0.00016174268422037812, + "loss": 0.3029, + "step": 36790 + }, + { + "epoch": 0.5740313221438822, + "grad_norm": 3.6200709342956543, + "learning_rate": 0.00016173228510222334, + "loss": 0.4174, + "step": 36800 + }, + { + "epoch": 0.574187308916204, + "grad_norm": 1.447937250137329, + "learning_rate": 0.00016172188598406856, + "loss": 0.1763, + "step": 36810 + }, + { + "epoch": 0.5743432956885256, + "grad_norm": 1.4020819664001465, + "learning_rate": 0.00016171148686591378, + "loss": 0.3994, + "step": 36820 + }, + { + "epoch": 0.5744992824608474, + "grad_norm": 1.4079807996749878, + "learning_rate": 0.000161701087747759, + "loss": 0.2051, + "step": 36830 + }, + { + "epoch": 0.574655269233169, + "grad_norm": 0.05268567427992821, + "learning_rate": 0.00016169068862960422, + "loss": 0.1773, + "step": 36840 + }, + { + "epoch": 0.5748112560054908, + "grad_norm": 2.248767614364624, + "learning_rate": 0.00016168028951144944, + "loss": 0.2534, + "step": 36850 + }, + { + "epoch": 0.5749672427778124, + "grad_norm": 0.3726302981376648, + "learning_rate": 0.00016166989039329466, + "loss": 0.328, + "step": 36860 + }, + { + "epoch": 0.5751232295501342, + "grad_norm": 1.0772110223770142, + "learning_rate": 0.00016165949127513988, + "loss": 0.2069, + "step": 36870 + }, + { + "epoch": 0.5752792163224558, + "grad_norm": 0.11339398473501205, + "learning_rate": 0.0001616490921569851, + "loss": 0.1827, + "step": 36880 + }, + { + "epoch": 0.5754352030947776, + "grad_norm": 1.2981103658676147, + "learning_rate": 0.00016163869303883032, + "loss": 0.1378, + "step": 36890 + }, + { + "epoch": 0.5755911898670992, + "grad_norm": 0.3302285075187683, + "learning_rate": 0.00016162829392067554, + "loss": 0.2311, + "step": 36900 + }, + { + "epoch": 0.575747176639421, + "grad_norm": 0.10168436169624329, + "learning_rate": 0.00016161789480252076, + "loss": 0.1678, + "step": 36910 + }, + { + "epoch": 0.5759031634117426, + "grad_norm": 1.6778357028961182, + "learning_rate": 0.00016160749568436598, + "loss": 0.2385, + "step": 36920 + }, + { + "epoch": 0.5760591501840644, + "grad_norm": 8.1710205078125, + "learning_rate": 0.0001615970965662112, + "loss": 0.3287, + "step": 36930 + }, + { + "epoch": 0.5762151369563862, + "grad_norm": 2.6481053829193115, + "learning_rate": 0.00016158669744805641, + "loss": 0.3119, + "step": 36940 + }, + { + "epoch": 0.5763711237287078, + "grad_norm": 0.3120105266571045, + "learning_rate": 0.00016157629832990163, + "loss": 0.1985, + "step": 36950 + }, + { + "epoch": 0.5765271105010296, + "grad_norm": 2.7221410274505615, + "learning_rate": 0.00016156589921174685, + "loss": 0.3697, + "step": 36960 + }, + { + "epoch": 0.5766830972733512, + "grad_norm": 0.34397706389427185, + "learning_rate": 0.00016155550009359207, + "loss": 0.2162, + "step": 36970 + }, + { + "epoch": 0.576839084045673, + "grad_norm": 4.6421003341674805, + "learning_rate": 0.0001615451009754373, + "loss": 0.2892, + "step": 36980 + }, + { + "epoch": 0.5769950708179946, + "grad_norm": 1.5395855903625488, + "learning_rate": 0.0001615347018572825, + "loss": 0.1258, + "step": 36990 + }, + { + "epoch": 0.5771510575903164, + "grad_norm": 0.4586782455444336, + "learning_rate": 0.00016152430273912773, + "loss": 0.1752, + "step": 37000 + }, + { + "epoch": 0.577307044362638, + "grad_norm": 0.4277658760547638, + "learning_rate": 0.00016151390362097295, + "loss": 0.1722, + "step": 37010 + }, + { + "epoch": 0.5774630311349598, + "grad_norm": 1.0282249450683594, + "learning_rate": 0.00016150350450281817, + "loss": 0.1361, + "step": 37020 + }, + { + "epoch": 0.5776190179072814, + "grad_norm": 0.2496921867132187, + "learning_rate": 0.0001614931053846634, + "loss": 0.0954, + "step": 37030 + }, + { + "epoch": 0.5777750046796032, + "grad_norm": 1.7900562286376953, + "learning_rate": 0.0001614827062665086, + "loss": 0.1284, + "step": 37040 + }, + { + "epoch": 0.5779309914519248, + "grad_norm": 0.6629091501235962, + "learning_rate": 0.00016147230714835383, + "loss": 0.2657, + "step": 37050 + }, + { + "epoch": 0.5780869782242466, + "grad_norm": 2.502882242202759, + "learning_rate": 0.00016146190803019905, + "loss": 0.207, + "step": 37060 + }, + { + "epoch": 0.5782429649965682, + "grad_norm": 1.486069917678833, + "learning_rate": 0.00016145150891204427, + "loss": 0.1232, + "step": 37070 + }, + { + "epoch": 0.57839895176889, + "grad_norm": 0.13543163239955902, + "learning_rate": 0.0001614411097938895, + "loss": 0.3245, + "step": 37080 + }, + { + "epoch": 0.5785549385412118, + "grad_norm": 3.9703898429870605, + "learning_rate": 0.0001614307106757347, + "loss": 0.3104, + "step": 37090 + }, + { + "epoch": 0.5787109253135334, + "grad_norm": 3.2236313819885254, + "learning_rate": 0.00016142031155757993, + "loss": 0.1739, + "step": 37100 + }, + { + "epoch": 0.5788669120858552, + "grad_norm": 0.3854866325855255, + "learning_rate": 0.00016140991243942515, + "loss": 0.084, + "step": 37110 + }, + { + "epoch": 0.5790228988581768, + "grad_norm": 0.1981169581413269, + "learning_rate": 0.00016139951332127037, + "loss": 0.2367, + "step": 37120 + }, + { + "epoch": 0.5791788856304986, + "grad_norm": 2.7315785884857178, + "learning_rate": 0.0001613891142031156, + "loss": 0.1816, + "step": 37130 + }, + { + "epoch": 0.5793348724028202, + "grad_norm": 0.48251378536224365, + "learning_rate": 0.0001613787150849608, + "loss": 0.2123, + "step": 37140 + }, + { + "epoch": 0.579490859175142, + "grad_norm": 5.0431413650512695, + "learning_rate": 0.00016136831596680603, + "loss": 0.2489, + "step": 37150 + }, + { + "epoch": 0.5796468459474636, + "grad_norm": 0.81981360912323, + "learning_rate": 0.00016135791684865125, + "loss": 0.5235, + "step": 37160 + }, + { + "epoch": 0.5798028327197854, + "grad_norm": 0.5701965689659119, + "learning_rate": 0.00016134751773049647, + "loss": 0.1709, + "step": 37170 + }, + { + "epoch": 0.579958819492107, + "grad_norm": 1.3905616998672485, + "learning_rate": 0.00016133711861234169, + "loss": 0.2641, + "step": 37180 + }, + { + "epoch": 0.5801148062644288, + "grad_norm": 2.281285047531128, + "learning_rate": 0.0001613267194941869, + "loss": 0.3915, + "step": 37190 + }, + { + "epoch": 0.5802707930367504, + "grad_norm": 1.2038309574127197, + "learning_rate": 0.00016131632037603213, + "loss": 0.0881, + "step": 37200 + }, + { + "epoch": 0.5804267798090722, + "grad_norm": 0.3096400201320648, + "learning_rate": 0.00016130592125787735, + "loss": 0.1341, + "step": 37210 + }, + { + "epoch": 0.5805827665813938, + "grad_norm": 0.10219905525445938, + "learning_rate": 0.00016129552213972256, + "loss": 0.0867, + "step": 37220 + }, + { + "epoch": 0.5807387533537156, + "grad_norm": 0.13869646191596985, + "learning_rate": 0.00016128512302156778, + "loss": 0.1621, + "step": 37230 + }, + { + "epoch": 0.5808947401260373, + "grad_norm": 0.7992938160896301, + "learning_rate": 0.000161274723903413, + "loss": 0.275, + "step": 37240 + }, + { + "epoch": 0.581050726898359, + "grad_norm": 0.2635972201824188, + "learning_rate": 0.00016126432478525822, + "loss": 0.2429, + "step": 37250 + }, + { + "epoch": 0.5812067136706808, + "grad_norm": 1.7211238145828247, + "learning_rate": 0.00016125392566710344, + "loss": 0.1541, + "step": 37260 + }, + { + "epoch": 0.5813627004430024, + "grad_norm": 2.2448506355285645, + "learning_rate": 0.00016124352654894866, + "loss": 0.3689, + "step": 37270 + }, + { + "epoch": 0.5815186872153242, + "grad_norm": 1.238708257675171, + "learning_rate": 0.00016123312743079388, + "loss": 0.3292, + "step": 37280 + }, + { + "epoch": 0.5816746739876458, + "grad_norm": 0.5670503377914429, + "learning_rate": 0.0001612227283126391, + "loss": 0.1151, + "step": 37290 + }, + { + "epoch": 0.5818306607599676, + "grad_norm": 2.287304401397705, + "learning_rate": 0.00016121232919448432, + "loss": 0.3558, + "step": 37300 + }, + { + "epoch": 0.5819866475322892, + "grad_norm": 0.8009629845619202, + "learning_rate": 0.00016120193007632954, + "loss": 0.1861, + "step": 37310 + }, + { + "epoch": 0.582142634304611, + "grad_norm": 1.1901466846466064, + "learning_rate": 0.00016119153095817476, + "loss": 0.089, + "step": 37320 + }, + { + "epoch": 0.5822986210769326, + "grad_norm": 1.7198141813278198, + "learning_rate": 0.00016118113184001998, + "loss": 0.2931, + "step": 37330 + }, + { + "epoch": 0.5824546078492544, + "grad_norm": 0.8832749724388123, + "learning_rate": 0.0001611707327218652, + "loss": 0.2342, + "step": 37340 + }, + { + "epoch": 0.582610594621576, + "grad_norm": 1.0623290538787842, + "learning_rate": 0.00016116033360371042, + "loss": 0.2317, + "step": 37350 + }, + { + "epoch": 0.5827665813938978, + "grad_norm": 0.5173699855804443, + "learning_rate": 0.00016114993448555564, + "loss": 0.1299, + "step": 37360 + }, + { + "epoch": 0.5829225681662195, + "grad_norm": 0.05118720978498459, + "learning_rate": 0.00016113953536740086, + "loss": 0.0786, + "step": 37370 + }, + { + "epoch": 0.5830785549385412, + "grad_norm": 2.0060107707977295, + "learning_rate": 0.00016112913624924608, + "loss": 0.266, + "step": 37380 + }, + { + "epoch": 0.5832345417108629, + "grad_norm": 5.121693134307861, + "learning_rate": 0.0001611187371310913, + "loss": 0.1824, + "step": 37390 + }, + { + "epoch": 0.5833905284831846, + "grad_norm": 0.3568089008331299, + "learning_rate": 0.00016110833801293652, + "loss": 0.3189, + "step": 37400 + }, + { + "epoch": 0.5835465152555064, + "grad_norm": 0.8631492853164673, + "learning_rate": 0.00016109793889478174, + "loss": 0.1724, + "step": 37410 + }, + { + "epoch": 0.583702502027828, + "grad_norm": 1.9214404821395874, + "learning_rate": 0.00016108753977662693, + "loss": 0.2094, + "step": 37420 + }, + { + "epoch": 0.5838584888001498, + "grad_norm": 1.9055646657943726, + "learning_rate": 0.00016107714065847218, + "loss": 0.1258, + "step": 37430 + }, + { + "epoch": 0.5840144755724714, + "grad_norm": 0.6614194512367249, + "learning_rate": 0.00016106674154031737, + "loss": 0.1966, + "step": 37440 + }, + { + "epoch": 0.5841704623447932, + "grad_norm": 0.286883682012558, + "learning_rate": 0.00016105634242216262, + "loss": 0.588, + "step": 37450 + }, + { + "epoch": 0.5843264491171148, + "grad_norm": 0.5599405765533447, + "learning_rate": 0.0001610459433040078, + "loss": 0.3414, + "step": 37460 + }, + { + "epoch": 0.5844824358894366, + "grad_norm": 1.5908915996551514, + "learning_rate": 0.00016103554418585306, + "loss": 0.1264, + "step": 37470 + }, + { + "epoch": 0.5846384226617583, + "grad_norm": 1.0729930400848389, + "learning_rate": 0.00016102514506769825, + "loss": 0.4815, + "step": 37480 + }, + { + "epoch": 0.58479440943408, + "grad_norm": 0.8912356495857239, + "learning_rate": 0.0001610147459495435, + "loss": 0.2092, + "step": 37490 + }, + { + "epoch": 0.5849503962064017, + "grad_norm": 0.24116197228431702, + "learning_rate": 0.0001610043468313887, + "loss": 0.1552, + "step": 37500 + }, + { + "epoch": 0.5851063829787234, + "grad_norm": 2.245192527770996, + "learning_rate": 0.00016099394771323393, + "loss": 0.0553, + "step": 37510 + }, + { + "epoch": 0.5852623697510451, + "grad_norm": 0.11538795381784439, + "learning_rate": 0.00016098354859507913, + "loss": 0.0889, + "step": 37520 + }, + { + "epoch": 0.5854183565233668, + "grad_norm": 1.0372689962387085, + "learning_rate": 0.00016097314947692437, + "loss": 0.3909, + "step": 37530 + }, + { + "epoch": 0.5855743432956885, + "grad_norm": 1.1946550607681274, + "learning_rate": 0.00016096275035876957, + "loss": 0.2167, + "step": 37540 + }, + { + "epoch": 0.5857303300680102, + "grad_norm": 1.3474591970443726, + "learning_rate": 0.0001609523512406148, + "loss": 0.1561, + "step": 37550 + }, + { + "epoch": 0.585886316840332, + "grad_norm": 0.008936616592109203, + "learning_rate": 0.00016094195212246, + "loss": 0.3991, + "step": 37560 + }, + { + "epoch": 0.5860423036126536, + "grad_norm": 1.8822320699691772, + "learning_rate": 0.00016093155300430525, + "loss": 0.2438, + "step": 37570 + }, + { + "epoch": 0.5861982903849754, + "grad_norm": 1.1707524061203003, + "learning_rate": 0.00016092115388615044, + "loss": 0.308, + "step": 37580 + }, + { + "epoch": 0.586354277157297, + "grad_norm": 1.679641842842102, + "learning_rate": 0.0001609107547679957, + "loss": 0.159, + "step": 37590 + }, + { + "epoch": 0.5865102639296188, + "grad_norm": 1.6885284185409546, + "learning_rate": 0.00016090035564984088, + "loss": 0.157, + "step": 37600 + }, + { + "epoch": 0.5866662507019405, + "grad_norm": 0.3876996636390686, + "learning_rate": 0.00016088995653168613, + "loss": 0.4067, + "step": 37610 + }, + { + "epoch": 0.5868222374742622, + "grad_norm": 1.387480616569519, + "learning_rate": 0.00016087955741353132, + "loss": 0.1898, + "step": 37620 + }, + { + "epoch": 0.5869782242465839, + "grad_norm": 0.03653861582279205, + "learning_rate": 0.00016086915829537657, + "loss": 0.1652, + "step": 37630 + }, + { + "epoch": 0.5871342110189056, + "grad_norm": 0.6258412003517151, + "learning_rate": 0.00016085875917722176, + "loss": 0.2558, + "step": 37640 + }, + { + "epoch": 0.5872901977912273, + "grad_norm": 0.10519164055585861, + "learning_rate": 0.000160848360059067, + "loss": 0.243, + "step": 37650 + }, + { + "epoch": 0.587446184563549, + "grad_norm": 0.7370131015777588, + "learning_rate": 0.0001608379609409122, + "loss": 0.2343, + "step": 37660 + }, + { + "epoch": 0.5876021713358707, + "grad_norm": 0.6884517073631287, + "learning_rate": 0.00016082756182275745, + "loss": 0.1847, + "step": 37670 + }, + { + "epoch": 0.5877581581081924, + "grad_norm": 0.6326120495796204, + "learning_rate": 0.00016081716270460264, + "loss": 0.2258, + "step": 37680 + }, + { + "epoch": 0.5879141448805141, + "grad_norm": 7.697889804840088, + "learning_rate": 0.0001608067635864479, + "loss": 0.2068, + "step": 37690 + }, + { + "epoch": 0.5880701316528358, + "grad_norm": 1.8753420114517212, + "learning_rate": 0.00016079636446829308, + "loss": 0.3087, + "step": 37700 + }, + { + "epoch": 0.5882261184251576, + "grad_norm": 0.8591625094413757, + "learning_rate": 0.00016078596535013833, + "loss": 0.1082, + "step": 37710 + }, + { + "epoch": 0.5883821051974792, + "grad_norm": 0.6124423146247864, + "learning_rate": 0.00016077556623198352, + "loss": 0.1536, + "step": 37720 + }, + { + "epoch": 0.588538091969801, + "grad_norm": 0.1400165557861328, + "learning_rate": 0.00016076516711382877, + "loss": 0.145, + "step": 37730 + }, + { + "epoch": 0.5886940787421227, + "grad_norm": 1.9879004955291748, + "learning_rate": 0.00016075476799567396, + "loss": 0.4986, + "step": 37740 + }, + { + "epoch": 0.5888500655144444, + "grad_norm": 0.08687327802181244, + "learning_rate": 0.0001607443688775192, + "loss": 0.3365, + "step": 37750 + }, + { + "epoch": 0.5890060522867661, + "grad_norm": 0.3442278206348419, + "learning_rate": 0.0001607339697593644, + "loss": 0.2274, + "step": 37760 + }, + { + "epoch": 0.5891620390590878, + "grad_norm": 0.06124915927648544, + "learning_rate": 0.00016072357064120965, + "loss": 0.1915, + "step": 37770 + }, + { + "epoch": 0.5893180258314095, + "grad_norm": 1.0996816158294678, + "learning_rate": 0.00016071317152305484, + "loss": 0.1573, + "step": 37780 + }, + { + "epoch": 0.5894740126037312, + "grad_norm": 0.14738436043262482, + "learning_rate": 0.00016070277240490008, + "loss": 0.2524, + "step": 37790 + }, + { + "epoch": 0.5896299993760529, + "grad_norm": 2.3905467987060547, + "learning_rate": 0.00016069237328674528, + "loss": 0.1838, + "step": 37800 + }, + { + "epoch": 0.5897859861483746, + "grad_norm": 2.0800178050994873, + "learning_rate": 0.00016068197416859052, + "loss": 0.3749, + "step": 37810 + }, + { + "epoch": 0.5899419729206963, + "grad_norm": 3.254204511642456, + "learning_rate": 0.00016067157505043574, + "loss": 0.3777, + "step": 37820 + }, + { + "epoch": 0.590097959693018, + "grad_norm": 2.8447892665863037, + "learning_rate": 0.00016066117593228096, + "loss": 0.5338, + "step": 37830 + }, + { + "epoch": 0.5902539464653397, + "grad_norm": 0.624733030796051, + "learning_rate": 0.00016065077681412618, + "loss": 0.2331, + "step": 37840 + }, + { + "epoch": 0.5904099332376614, + "grad_norm": 0.6863507628440857, + "learning_rate": 0.0001606403776959714, + "loss": 0.1747, + "step": 37850 + }, + { + "epoch": 0.5905659200099832, + "grad_norm": 0.6785943508148193, + "learning_rate": 0.00016062997857781662, + "loss": 0.2064, + "step": 37860 + }, + { + "epoch": 0.5907219067823049, + "grad_norm": 0.5739938616752625, + "learning_rate": 0.00016061957945966181, + "loss": 0.2368, + "step": 37870 + }, + { + "epoch": 0.5908778935546266, + "grad_norm": 0.0616387277841568, + "learning_rate": 0.00016060918034150706, + "loss": 0.2224, + "step": 37880 + }, + { + "epoch": 0.5910338803269483, + "grad_norm": 2.745957136154175, + "learning_rate": 0.00016059878122335225, + "loss": 0.3161, + "step": 37890 + }, + { + "epoch": 0.59118986709927, + "grad_norm": 2.52472186088562, + "learning_rate": 0.0001605883821051975, + "loss": 0.1713, + "step": 37900 + }, + { + "epoch": 0.5913458538715917, + "grad_norm": 1.5691877603530884, + "learning_rate": 0.0001605779829870427, + "loss": 0.1402, + "step": 37910 + }, + { + "epoch": 0.5915018406439134, + "grad_norm": 1.848261833190918, + "learning_rate": 0.00016056758386888794, + "loss": 0.2305, + "step": 37920 + }, + { + "epoch": 0.5916578274162351, + "grad_norm": 0.36622482538223267, + "learning_rate": 0.00016055718475073313, + "loss": 0.1834, + "step": 37930 + }, + { + "epoch": 0.5918138141885568, + "grad_norm": 0.21201446652412415, + "learning_rate": 0.00016054678563257838, + "loss": 0.4682, + "step": 37940 + }, + { + "epoch": 0.5919698009608785, + "grad_norm": 0.38921502232551575, + "learning_rate": 0.00016053638651442357, + "loss": 0.2892, + "step": 37950 + }, + { + "epoch": 0.5921257877332002, + "grad_norm": 0.20777840912342072, + "learning_rate": 0.00016052598739626882, + "loss": 0.104, + "step": 37960 + }, + { + "epoch": 0.5922817745055219, + "grad_norm": 0.7090150713920593, + "learning_rate": 0.000160515588278114, + "loss": 0.2592, + "step": 37970 + }, + { + "epoch": 0.5924377612778436, + "grad_norm": 1.9939709901809692, + "learning_rate": 0.00016050518915995926, + "loss": 0.247, + "step": 37980 + }, + { + "epoch": 0.5925937480501653, + "grad_norm": 0.8696603178977966, + "learning_rate": 0.00016049479004180445, + "loss": 0.301, + "step": 37990 + }, + { + "epoch": 0.592749734822487, + "grad_norm": 0.4119095206260681, + "learning_rate": 0.0001604843909236497, + "loss": 0.2011, + "step": 38000 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.314039984128e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5ce841738b8b0e9001b883e0e34560afec1f2c00 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56489c9494f7cfb591e72d95274017a007010a85ba2cef11c52eb3ee1c762037 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..139abb604d72fae64b9c59b73f34f1bb4dca54f7 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:06a112dfa20b3bde41dcf08bbeaf90f2699070899fbead9c1f1eda0921af1acf +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..be4dcbaee80884cacc04302a41a0c8e5aa92e385 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..b306dac88c4155dd0c8dde48361e96ae0d98235e Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..7bcd0c4e42725e3fbb8c424615a1b9cca304159e --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/trainer_state.json @@ -0,0 +1,26984 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.6005490734385724, + "eval_steps": 500, + "global_step": 38500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + }, + { + "epoch": 0.5071129968178698, + "grad_norm": 1.0075199604034424, + "learning_rate": 0.00016619350679062414, + "loss": 0.2569, + "step": 32510 + }, + { + "epoch": 0.5072689835901916, + "grad_norm": 1.3489066362380981, + "learning_rate": 0.0001661831076724694, + "loss": 0.2565, + "step": 32520 + }, + { + "epoch": 0.5074249703625132, + "grad_norm": 0.2534058094024658, + "learning_rate": 0.00016617270855431458, + "loss": 0.2292, + "step": 32530 + }, + { + "epoch": 0.507580957134835, + "grad_norm": 0.31723281741142273, + "learning_rate": 0.00016616230943615983, + "loss": 0.1467, + "step": 32540 + }, + { + "epoch": 0.5077369439071566, + "grad_norm": 0.29659849405288696, + "learning_rate": 0.00016615191031800502, + "loss": 0.2768, + "step": 32550 + }, + { + "epoch": 0.5078929306794784, + "grad_norm": 1.5246435403823853, + "learning_rate": 0.00016614151119985027, + "loss": 0.216, + "step": 32560 + }, + { + "epoch": 0.5080489174518001, + "grad_norm": 0.6026854515075684, + "learning_rate": 0.00016613111208169546, + "loss": 0.2277, + "step": 32570 + }, + { + "epoch": 0.5082049042241218, + "grad_norm": 0.41994890570640564, + "learning_rate": 0.0001661207129635407, + "loss": 0.2433, + "step": 32580 + }, + { + "epoch": 0.5083608909964435, + "grad_norm": 2.084599733352661, + "learning_rate": 0.00016611031384538593, + "loss": 0.4081, + "step": 32590 + }, + { + "epoch": 0.5085168777687652, + "grad_norm": 0.2945229709148407, + "learning_rate": 0.00016609991472723115, + "loss": 0.1329, + "step": 32600 + }, + { + "epoch": 0.508672864541087, + "grad_norm": 1.4981423616409302, + "learning_rate": 0.00016608951560907636, + "loss": 0.2855, + "step": 32610 + }, + { + "epoch": 0.5088288513134086, + "grad_norm": 0.961737871170044, + "learning_rate": 0.00016607911649092158, + "loss": 0.1654, + "step": 32620 + }, + { + "epoch": 0.5089848380857304, + "grad_norm": 1.49734628200531, + "learning_rate": 0.0001660687173727668, + "loss": 0.2081, + "step": 32630 + }, + { + "epoch": 0.509140824858052, + "grad_norm": 2.538135051727295, + "learning_rate": 0.00016605831825461202, + "loss": 0.3023, + "step": 32640 + }, + { + "epoch": 0.5092968116303738, + "grad_norm": 0.2981272041797638, + "learning_rate": 0.00016604791913645724, + "loss": 0.214, + "step": 32650 + }, + { + "epoch": 0.5094527984026954, + "grad_norm": 1.8863650560379028, + "learning_rate": 0.00016603752001830246, + "loss": 0.3363, + "step": 32660 + }, + { + "epoch": 0.5096087851750172, + "grad_norm": 0.022720983251929283, + "learning_rate": 0.00016602712090014768, + "loss": 0.2463, + "step": 32670 + }, + { + "epoch": 0.5097647719473388, + "grad_norm": 1.154977798461914, + "learning_rate": 0.0001660167217819929, + "loss": 0.4484, + "step": 32680 + }, + { + "epoch": 0.5099207587196606, + "grad_norm": 0.9470499157905579, + "learning_rate": 0.00016600632266383812, + "loss": 0.1196, + "step": 32690 + }, + { + "epoch": 0.5100767454919822, + "grad_norm": 0.19955401122570038, + "learning_rate": 0.00016599592354568334, + "loss": 0.1762, + "step": 32700 + }, + { + "epoch": 0.510232732264304, + "grad_norm": 0.4098282754421234, + "learning_rate": 0.00016598552442752856, + "loss": 0.2739, + "step": 32710 + }, + { + "epoch": 0.5103887190366257, + "grad_norm": 0.21697258949279785, + "learning_rate": 0.00016597512530937378, + "loss": 0.2555, + "step": 32720 + }, + { + "epoch": 0.5105447058089474, + "grad_norm": 4.894479751586914, + "learning_rate": 0.000165964726191219, + "loss": 0.3137, + "step": 32730 + }, + { + "epoch": 0.5107006925812692, + "grad_norm": 0.8396711349487305, + "learning_rate": 0.00016595432707306422, + "loss": 0.3105, + "step": 32740 + }, + { + "epoch": 0.5108566793535908, + "grad_norm": 1.5494674444198608, + "learning_rate": 0.00016594392795490944, + "loss": 0.2271, + "step": 32750 + }, + { + "epoch": 0.5110126661259126, + "grad_norm": 1.3409888744354248, + "learning_rate": 0.00016593352883675466, + "loss": 0.1796, + "step": 32760 + }, + { + "epoch": 0.5111686528982342, + "grad_norm": 1.7590911388397217, + "learning_rate": 0.00016592312971859988, + "loss": 0.2185, + "step": 32770 + }, + { + "epoch": 0.511324639670556, + "grad_norm": 0.7350002527236938, + "learning_rate": 0.0001659127306004451, + "loss": 0.1911, + "step": 32780 + }, + { + "epoch": 0.5114806264428776, + "grad_norm": 0.7176411747932434, + "learning_rate": 0.00016590233148229032, + "loss": 0.6918, + "step": 32790 + }, + { + "epoch": 0.5116366132151994, + "grad_norm": 2.5005321502685547, + "learning_rate": 0.00016589193236413554, + "loss": 0.2329, + "step": 32800 + }, + { + "epoch": 0.511792599987521, + "grad_norm": 1.3916250467300415, + "learning_rate": 0.00016588153324598076, + "loss": 0.2102, + "step": 32810 + }, + { + "epoch": 0.5119485867598428, + "grad_norm": 0.721671998500824, + "learning_rate": 0.00016587113412782598, + "loss": 0.2965, + "step": 32820 + }, + { + "epoch": 0.5121045735321644, + "grad_norm": 0.6120114326477051, + "learning_rate": 0.0001658607350096712, + "loss": 0.2661, + "step": 32830 + }, + { + "epoch": 0.5122605603044862, + "grad_norm": 0.29123347997665405, + "learning_rate": 0.00016585033589151642, + "loss": 0.2455, + "step": 32840 + }, + { + "epoch": 0.5124165470768078, + "grad_norm": 2.7300868034362793, + "learning_rate": 0.00016583993677336164, + "loss": 0.2049, + "step": 32850 + }, + { + "epoch": 0.5125725338491296, + "grad_norm": 1.801928997039795, + "learning_rate": 0.00016582953765520686, + "loss": 0.1725, + "step": 32860 + }, + { + "epoch": 0.5127285206214514, + "grad_norm": 0.6406471729278564, + "learning_rate": 0.00016581913853705208, + "loss": 0.164, + "step": 32870 + }, + { + "epoch": 0.512884507393773, + "grad_norm": 0.9627325534820557, + "learning_rate": 0.0001658087394188973, + "loss": 0.1728, + "step": 32880 + }, + { + "epoch": 0.5130404941660948, + "grad_norm": 1.5446962118148804, + "learning_rate": 0.00016579834030074251, + "loss": 0.2345, + "step": 32890 + }, + { + "epoch": 0.5131964809384164, + "grad_norm": 0.28092706203460693, + "learning_rate": 0.0001657879411825877, + "loss": 0.2124, + "step": 32900 + }, + { + "epoch": 0.5133524677107382, + "grad_norm": 1.7751421928405762, + "learning_rate": 0.00016577754206443295, + "loss": 0.2367, + "step": 32910 + }, + { + "epoch": 0.5135084544830598, + "grad_norm": 1.6964634656906128, + "learning_rate": 0.00016576714294627815, + "loss": 0.2361, + "step": 32920 + }, + { + "epoch": 0.5136644412553816, + "grad_norm": 0.06759823113679886, + "learning_rate": 0.0001657567438281234, + "loss": 0.2329, + "step": 32930 + }, + { + "epoch": 0.5138204280277032, + "grad_norm": 2.2519280910491943, + "learning_rate": 0.00016574634470996859, + "loss": 0.1079, + "step": 32940 + }, + { + "epoch": 0.513976414800025, + "grad_norm": 3.167576551437378, + "learning_rate": 0.00016573594559181383, + "loss": 0.1764, + "step": 32950 + }, + { + "epoch": 0.5141324015723466, + "grad_norm": 1.6734397411346436, + "learning_rate": 0.00016572554647365903, + "loss": 0.2878, + "step": 32960 + }, + { + "epoch": 0.5142883883446684, + "grad_norm": 0.35627374053001404, + "learning_rate": 0.00016571514735550427, + "loss": 0.2337, + "step": 32970 + }, + { + "epoch": 0.51444437511699, + "grad_norm": 1.6321039199829102, + "learning_rate": 0.00016570474823734946, + "loss": 0.2551, + "step": 32980 + }, + { + "epoch": 0.5146003618893118, + "grad_norm": 1.352982997894287, + "learning_rate": 0.0001656943491191947, + "loss": 0.4467, + "step": 32990 + }, + { + "epoch": 0.5147563486616334, + "grad_norm": 0.3558129072189331, + "learning_rate": 0.0001656839500010399, + "loss": 0.1913, + "step": 33000 + }, + { + "epoch": 0.5149123354339552, + "grad_norm": 0.041972286999225616, + "learning_rate": 0.00016567355088288515, + "loss": 0.1704, + "step": 33010 + }, + { + "epoch": 0.515068322206277, + "grad_norm": 1.4168939590454102, + "learning_rate": 0.00016566315176473034, + "loss": 0.4866, + "step": 33020 + }, + { + "epoch": 0.5152243089785986, + "grad_norm": 3.0042929649353027, + "learning_rate": 0.0001656527526465756, + "loss": 0.2473, + "step": 33030 + }, + { + "epoch": 0.5153802957509204, + "grad_norm": 1.027274489402771, + "learning_rate": 0.00016564235352842078, + "loss": 0.436, + "step": 33040 + }, + { + "epoch": 0.515536282523242, + "grad_norm": 1.2330195903778076, + "learning_rate": 0.00016563195441026603, + "loss": 0.4615, + "step": 33050 + }, + { + "epoch": 0.5156922692955638, + "grad_norm": 2.7145066261291504, + "learning_rate": 0.00016562155529211122, + "loss": 0.278, + "step": 33060 + }, + { + "epoch": 0.5158482560678854, + "grad_norm": 1.1273893117904663, + "learning_rate": 0.00016561115617395647, + "loss": 0.3767, + "step": 33070 + }, + { + "epoch": 0.5160042428402072, + "grad_norm": 2.466265916824341, + "learning_rate": 0.00016560075705580166, + "loss": 0.1879, + "step": 33080 + }, + { + "epoch": 0.5161602296125288, + "grad_norm": 0.3759515583515167, + "learning_rate": 0.0001655903579376469, + "loss": 0.4004, + "step": 33090 + }, + { + "epoch": 0.5163162163848506, + "grad_norm": 1.8821336030960083, + "learning_rate": 0.0001655799588194921, + "loss": 0.1622, + "step": 33100 + }, + { + "epoch": 0.5164722031571722, + "grad_norm": 0.8162096738815308, + "learning_rate": 0.00016556955970133735, + "loss": 0.1341, + "step": 33110 + }, + { + "epoch": 0.516628189929494, + "grad_norm": 0.008498801849782467, + "learning_rate": 0.00016555916058318254, + "loss": 0.2546, + "step": 33120 + }, + { + "epoch": 0.5167841767018156, + "grad_norm": 0.17536675930023193, + "learning_rate": 0.00016554876146502779, + "loss": 0.2934, + "step": 33130 + }, + { + "epoch": 0.5169401634741374, + "grad_norm": 3.539898633956909, + "learning_rate": 0.00016553836234687298, + "loss": 0.4959, + "step": 33140 + }, + { + "epoch": 0.517096150246459, + "grad_norm": 2.190887212753296, + "learning_rate": 0.00016552796322871823, + "loss": 0.2914, + "step": 33150 + }, + { + "epoch": 0.5172521370187808, + "grad_norm": 1.6388943195343018, + "learning_rate": 0.00016551756411056342, + "loss": 0.1317, + "step": 33160 + }, + { + "epoch": 0.5174081237911026, + "grad_norm": 0.33063387870788574, + "learning_rate": 0.00016550716499240866, + "loss": 0.4258, + "step": 33170 + }, + { + "epoch": 0.5175641105634242, + "grad_norm": 0.5168288946151733, + "learning_rate": 0.00016549676587425386, + "loss": 0.2548, + "step": 33180 + }, + { + "epoch": 0.517720097335746, + "grad_norm": 3.9570281505584717, + "learning_rate": 0.0001654863667560991, + "loss": 0.3179, + "step": 33190 + }, + { + "epoch": 0.5178760841080676, + "grad_norm": 1.127255916595459, + "learning_rate": 0.0001654759676379443, + "loss": 0.233, + "step": 33200 + }, + { + "epoch": 0.5180320708803894, + "grad_norm": 0.36299410462379456, + "learning_rate": 0.00016546556851978954, + "loss": 0.4614, + "step": 33210 + }, + { + "epoch": 0.518188057652711, + "grad_norm": 2.914668321609497, + "learning_rate": 0.00016545516940163474, + "loss": 0.1603, + "step": 33220 + }, + { + "epoch": 0.5183440444250328, + "grad_norm": 6.991094589233398, + "learning_rate": 0.00016544477028347998, + "loss": 0.2259, + "step": 33230 + }, + { + "epoch": 0.5185000311973544, + "grad_norm": 2.0594241619110107, + "learning_rate": 0.00016543437116532517, + "loss": 0.1885, + "step": 33240 + }, + { + "epoch": 0.5186560179696762, + "grad_norm": 0.5594693422317505, + "learning_rate": 0.00016542397204717042, + "loss": 0.4333, + "step": 33250 + }, + { + "epoch": 0.5188120047419978, + "grad_norm": 4.314854621887207, + "learning_rate": 0.00016541357292901561, + "loss": 0.2103, + "step": 33260 + }, + { + "epoch": 0.5189679915143196, + "grad_norm": 0.5609000325202942, + "learning_rate": 0.00016540317381086086, + "loss": 0.2498, + "step": 33270 + }, + { + "epoch": 0.5191239782866413, + "grad_norm": 3.680088758468628, + "learning_rate": 0.00016539277469270605, + "loss": 0.3465, + "step": 33280 + }, + { + "epoch": 0.519279965058963, + "grad_norm": 1.4993672370910645, + "learning_rate": 0.0001653823755745513, + "loss": 0.3272, + "step": 33290 + }, + { + "epoch": 0.5194359518312847, + "grad_norm": 0.18005084991455078, + "learning_rate": 0.0001653719764563965, + "loss": 0.1706, + "step": 33300 + }, + { + "epoch": 0.5195919386036064, + "grad_norm": 0.5105524659156799, + "learning_rate": 0.00016536157733824174, + "loss": 0.0937, + "step": 33310 + }, + { + "epoch": 0.5197479253759281, + "grad_norm": 1.2806589603424072, + "learning_rate": 0.00016535117822008693, + "loss": 0.1578, + "step": 33320 + }, + { + "epoch": 0.5199039121482498, + "grad_norm": 0.22015990316867828, + "learning_rate": 0.00016534077910193218, + "loss": 0.2081, + "step": 33330 + }, + { + "epoch": 0.5200598989205716, + "grad_norm": 0.34638991951942444, + "learning_rate": 0.00016533037998377737, + "loss": 0.2389, + "step": 33340 + }, + { + "epoch": 0.5202158856928932, + "grad_norm": 0.8709302544593811, + "learning_rate": 0.0001653199808656226, + "loss": 0.2112, + "step": 33350 + }, + { + "epoch": 0.520371872465215, + "grad_norm": 1.2309521436691284, + "learning_rate": 0.0001653095817474678, + "loss": 0.3448, + "step": 33360 + }, + { + "epoch": 0.5205278592375366, + "grad_norm": 1.3155115842819214, + "learning_rate": 0.00016529918262931303, + "loss": 0.1572, + "step": 33370 + }, + { + "epoch": 0.5206838460098584, + "grad_norm": 0.358751118183136, + "learning_rate": 0.00016528878351115825, + "loss": 0.3011, + "step": 33380 + }, + { + "epoch": 0.52083983278218, + "grad_norm": 3.800628662109375, + "learning_rate": 0.00016527838439300347, + "loss": 0.458, + "step": 33390 + }, + { + "epoch": 0.5209958195545018, + "grad_norm": 1.1253124475479126, + "learning_rate": 0.0001652679852748487, + "loss": 0.1238, + "step": 33400 + }, + { + "epoch": 0.5211518063268235, + "grad_norm": 0.46979108452796936, + "learning_rate": 0.0001652575861566939, + "loss": 0.0745, + "step": 33410 + }, + { + "epoch": 0.5213077930991452, + "grad_norm": 7.921492576599121, + "learning_rate": 0.00016524718703853913, + "loss": 0.4004, + "step": 33420 + }, + { + "epoch": 0.5214637798714669, + "grad_norm": 1.643127679824829, + "learning_rate": 0.00016523678792038435, + "loss": 0.2858, + "step": 33430 + }, + { + "epoch": 0.5216197666437886, + "grad_norm": 7.096631050109863, + "learning_rate": 0.0001652263888022296, + "loss": 0.4735, + "step": 33440 + }, + { + "epoch": 0.5217757534161103, + "grad_norm": 0.7280219197273254, + "learning_rate": 0.0001652159896840748, + "loss": 0.1841, + "step": 33450 + }, + { + "epoch": 0.521931740188432, + "grad_norm": 0.9415522813796997, + "learning_rate": 0.00016520559056592003, + "loss": 0.2169, + "step": 33460 + }, + { + "epoch": 0.5220877269607537, + "grad_norm": 2.264862060546875, + "learning_rate": 0.00016519519144776523, + "loss": 0.1902, + "step": 33470 + }, + { + "epoch": 0.5222437137330754, + "grad_norm": 2.4566822052001953, + "learning_rate": 0.00016518479232961047, + "loss": 0.3329, + "step": 33480 + }, + { + "epoch": 0.5223997005053972, + "grad_norm": 0.7330378890037537, + "learning_rate": 0.00016517439321145567, + "loss": 0.2341, + "step": 33490 + }, + { + "epoch": 0.5225556872777188, + "grad_norm": 1.6409542560577393, + "learning_rate": 0.0001651639940933009, + "loss": 0.3077, + "step": 33500 + }, + { + "epoch": 0.5227116740500406, + "grad_norm": 1.0647432804107666, + "learning_rate": 0.0001651535949751461, + "loss": 0.1935, + "step": 33510 + }, + { + "epoch": 0.5228676608223622, + "grad_norm": 0.507420539855957, + "learning_rate": 0.00016514319585699135, + "loss": 0.1314, + "step": 33520 + }, + { + "epoch": 0.523023647594684, + "grad_norm": 1.0064164400100708, + "learning_rate": 0.00016513279673883654, + "loss": 0.344, + "step": 33530 + }, + { + "epoch": 0.5231796343670057, + "grad_norm": 1.1936209201812744, + "learning_rate": 0.0001651223976206818, + "loss": 0.1948, + "step": 33540 + }, + { + "epoch": 0.5233356211393274, + "grad_norm": 0.04992926865816116, + "learning_rate": 0.00016511199850252698, + "loss": 0.1363, + "step": 33550 + }, + { + "epoch": 0.5234916079116491, + "grad_norm": 1.8256275653839111, + "learning_rate": 0.00016510159938437223, + "loss": 0.443, + "step": 33560 + }, + { + "epoch": 0.5236475946839708, + "grad_norm": 4.914908409118652, + "learning_rate": 0.00016509120026621742, + "loss": 0.4654, + "step": 33570 + }, + { + "epoch": 0.5238035814562925, + "grad_norm": 1.1625486612319946, + "learning_rate": 0.00016508080114806267, + "loss": 0.1974, + "step": 33580 + }, + { + "epoch": 0.5239595682286142, + "grad_norm": 6.6192522048950195, + "learning_rate": 0.00016507040202990786, + "loss": 0.2131, + "step": 33590 + }, + { + "epoch": 0.5241155550009359, + "grad_norm": 0.7369065880775452, + "learning_rate": 0.0001650600029117531, + "loss": 0.2371, + "step": 33600 + }, + { + "epoch": 0.5242715417732576, + "grad_norm": 1.5238152742385864, + "learning_rate": 0.0001650496037935983, + "loss": 0.1557, + "step": 33610 + }, + { + "epoch": 0.5244275285455793, + "grad_norm": 1.0418007373809814, + "learning_rate": 0.00016503920467544355, + "loss": 0.1878, + "step": 33620 + }, + { + "epoch": 0.524583515317901, + "grad_norm": 0.790117084980011, + "learning_rate": 0.00016502880555728874, + "loss": 0.2195, + "step": 33630 + }, + { + "epoch": 0.5247395020902228, + "grad_norm": 1.6712257862091064, + "learning_rate": 0.000165018406439134, + "loss": 0.1602, + "step": 33640 + }, + { + "epoch": 0.5248954888625444, + "grad_norm": 0.19236230850219727, + "learning_rate": 0.00016500800732097918, + "loss": 0.2526, + "step": 33650 + }, + { + "epoch": 0.5250514756348662, + "grad_norm": 1.3519701957702637, + "learning_rate": 0.00016499760820282443, + "loss": 0.4686, + "step": 33660 + }, + { + "epoch": 0.5252074624071879, + "grad_norm": 1.694342851638794, + "learning_rate": 0.00016498720908466962, + "loss": 0.1859, + "step": 33670 + }, + { + "epoch": 0.5253634491795096, + "grad_norm": 5.225239276885986, + "learning_rate": 0.00016497680996651487, + "loss": 0.2944, + "step": 33680 + }, + { + "epoch": 0.5255194359518313, + "grad_norm": 2.0208842754364014, + "learning_rate": 0.00016496641084836006, + "loss": 0.2421, + "step": 33690 + }, + { + "epoch": 0.525675422724153, + "grad_norm": 0.7954996228218079, + "learning_rate": 0.0001649560117302053, + "loss": 0.2853, + "step": 33700 + }, + { + "epoch": 0.5258314094964747, + "grad_norm": 2.296086072921753, + "learning_rate": 0.0001649456126120505, + "loss": 0.1333, + "step": 33710 + }, + { + "epoch": 0.5259873962687964, + "grad_norm": 1.1779128313064575, + "learning_rate": 0.00016493521349389574, + "loss": 0.1066, + "step": 33720 + }, + { + "epoch": 0.5261433830411181, + "grad_norm": 0.1756065934896469, + "learning_rate": 0.00016492481437574094, + "loss": 0.1352, + "step": 33730 + }, + { + "epoch": 0.5262993698134398, + "grad_norm": 0.13100725412368774, + "learning_rate": 0.00016491441525758618, + "loss": 0.2399, + "step": 33740 + }, + { + "epoch": 0.5264553565857615, + "grad_norm": 5.532008171081543, + "learning_rate": 0.00016490401613943138, + "loss": 0.2896, + "step": 33750 + }, + { + "epoch": 0.5266113433580832, + "grad_norm": 1.319886565208435, + "learning_rate": 0.00016489361702127662, + "loss": 0.3275, + "step": 33760 + }, + { + "epoch": 0.5267673301304049, + "grad_norm": 1.5550974607467651, + "learning_rate": 0.00016488321790312182, + "loss": 0.2677, + "step": 33770 + }, + { + "epoch": 0.5269233169027266, + "grad_norm": 1.8936737775802612, + "learning_rate": 0.00016487281878496706, + "loss": 0.1955, + "step": 33780 + }, + { + "epoch": 0.5270793036750484, + "grad_norm": 0.3653401732444763, + "learning_rate": 0.00016486241966681226, + "loss": 0.0723, + "step": 33790 + }, + { + "epoch": 0.52723529044737, + "grad_norm": 2.861341714859009, + "learning_rate": 0.00016485202054865747, + "loss": 0.2412, + "step": 33800 + }, + { + "epoch": 0.5273912772196918, + "grad_norm": 1.5291428565979004, + "learning_rate": 0.0001648416214305027, + "loss": 0.0871, + "step": 33810 + }, + { + "epoch": 0.5275472639920135, + "grad_norm": 1.0372581481933594, + "learning_rate": 0.00016483122231234791, + "loss": 0.4705, + "step": 33820 + }, + { + "epoch": 0.5277032507643352, + "grad_norm": 1.1943141222000122, + "learning_rate": 0.00016482082319419313, + "loss": 0.2848, + "step": 33830 + }, + { + "epoch": 0.5278592375366569, + "grad_norm": 1.9008225202560425, + "learning_rate": 0.00016481042407603835, + "loss": 0.139, + "step": 33840 + }, + { + "epoch": 0.5280152243089786, + "grad_norm": 2.132089138031006, + "learning_rate": 0.00016480002495788357, + "loss": 0.2119, + "step": 33850 + }, + { + "epoch": 0.5281712110813003, + "grad_norm": 0.24524426460266113, + "learning_rate": 0.0001647896258397288, + "loss": 0.1643, + "step": 33860 + }, + { + "epoch": 0.528327197853622, + "grad_norm": 1.6469637155532837, + "learning_rate": 0.000164779226721574, + "loss": 0.363, + "step": 33870 + }, + { + "epoch": 0.5284831846259437, + "grad_norm": 0.8767328858375549, + "learning_rate": 0.00016476882760341923, + "loss": 0.2632, + "step": 33880 + }, + { + "epoch": 0.5286391713982654, + "grad_norm": 0.06347586214542389, + "learning_rate": 0.00016475842848526445, + "loss": 0.3204, + "step": 33890 + }, + { + "epoch": 0.5287951581705871, + "grad_norm": 0.09782540798187256, + "learning_rate": 0.00016474802936710967, + "loss": 0.2211, + "step": 33900 + }, + { + "epoch": 0.5289511449429088, + "grad_norm": 3.2998859882354736, + "learning_rate": 0.0001647376302489549, + "loss": 0.265, + "step": 33910 + }, + { + "epoch": 0.5291071317152305, + "grad_norm": 0.43594226241111755, + "learning_rate": 0.0001647272311308001, + "loss": 0.194, + "step": 33920 + }, + { + "epoch": 0.5292631184875523, + "grad_norm": 1.5166605710983276, + "learning_rate": 0.00016471683201264533, + "loss": 0.2675, + "step": 33930 + }, + { + "epoch": 0.529419105259874, + "grad_norm": 0.6056640148162842, + "learning_rate": 0.00016470643289449055, + "loss": 0.0576, + "step": 33940 + }, + { + "epoch": 0.5295750920321957, + "grad_norm": 0.25410348176956177, + "learning_rate": 0.00016469603377633577, + "loss": 0.1124, + "step": 33950 + }, + { + "epoch": 0.5297310788045174, + "grad_norm": 1.770642876625061, + "learning_rate": 0.000164685634658181, + "loss": 0.3295, + "step": 33960 + }, + { + "epoch": 0.5298870655768391, + "grad_norm": 0.0607205331325531, + "learning_rate": 0.0001646752355400262, + "loss": 0.2369, + "step": 33970 + }, + { + "epoch": 0.5300430523491608, + "grad_norm": 0.5557095408439636, + "learning_rate": 0.00016466483642187143, + "loss": 0.1681, + "step": 33980 + }, + { + "epoch": 0.5301990391214825, + "grad_norm": 0.5192957520484924, + "learning_rate": 0.00016465443730371665, + "loss": 0.2649, + "step": 33990 + }, + { + "epoch": 0.5303550258938042, + "grad_norm": 0.04804835096001625, + "learning_rate": 0.00016464403818556187, + "loss": 0.2639, + "step": 34000 + }, + { + "epoch": 0.5305110126661259, + "grad_norm": 0.02673129364848137, + "learning_rate": 0.0001646336390674071, + "loss": 0.169, + "step": 34010 + }, + { + "epoch": 0.5306669994384476, + "grad_norm": 1.0084244012832642, + "learning_rate": 0.0001646232399492523, + "loss": 0.353, + "step": 34020 + }, + { + "epoch": 0.5308229862107693, + "grad_norm": 2.2202091217041016, + "learning_rate": 0.00016461284083109753, + "loss": 0.3206, + "step": 34030 + }, + { + "epoch": 0.530978972983091, + "grad_norm": 0.5573744773864746, + "learning_rate": 0.00016460244171294275, + "loss": 0.231, + "step": 34040 + }, + { + "epoch": 0.5311349597554127, + "grad_norm": 0.39700084924697876, + "learning_rate": 0.00016459204259478797, + "loss": 0.1412, + "step": 34050 + }, + { + "epoch": 0.5312909465277345, + "grad_norm": 2.582963228225708, + "learning_rate": 0.00016458164347663319, + "loss": 0.3334, + "step": 34060 + }, + { + "epoch": 0.5314469333000561, + "grad_norm": 0.22781169414520264, + "learning_rate": 0.0001645712443584784, + "loss": 0.2767, + "step": 34070 + }, + { + "epoch": 0.5316029200723779, + "grad_norm": 1.2051042318344116, + "learning_rate": 0.00016456084524032362, + "loss": 0.2045, + "step": 34080 + }, + { + "epoch": 0.5317589068446996, + "grad_norm": 0.42760100960731506, + "learning_rate": 0.00016455044612216884, + "loss": 0.1151, + "step": 34090 + }, + { + "epoch": 0.5319148936170213, + "grad_norm": 0.14440476894378662, + "learning_rate": 0.00016454004700401406, + "loss": 0.2386, + "step": 34100 + }, + { + "epoch": 0.532070880389343, + "grad_norm": 2.2777981758117676, + "learning_rate": 0.00016452964788585928, + "loss": 0.2009, + "step": 34110 + }, + { + "epoch": 0.5322268671616647, + "grad_norm": 0.9206979274749756, + "learning_rate": 0.0001645192487677045, + "loss": 0.2745, + "step": 34120 + }, + { + "epoch": 0.5323828539339864, + "grad_norm": 1.6947574615478516, + "learning_rate": 0.00016450884964954972, + "loss": 0.2584, + "step": 34130 + }, + { + "epoch": 0.5325388407063081, + "grad_norm": 0.401444673538208, + "learning_rate": 0.00016449845053139494, + "loss": 0.2218, + "step": 34140 + }, + { + "epoch": 0.5326948274786298, + "grad_norm": 0.08261553198099136, + "learning_rate": 0.00016448805141324016, + "loss": 0.2775, + "step": 34150 + }, + { + "epoch": 0.5328508142509515, + "grad_norm": 0.1017974391579628, + "learning_rate": 0.00016447765229508538, + "loss": 0.2095, + "step": 34160 + }, + { + "epoch": 0.5330068010232732, + "grad_norm": 1.3759571313858032, + "learning_rate": 0.0001644672531769306, + "loss": 0.2643, + "step": 34170 + }, + { + "epoch": 0.5331627877955949, + "grad_norm": 1.2654389142990112, + "learning_rate": 0.00016445685405877582, + "loss": 0.2949, + "step": 34180 + }, + { + "epoch": 0.5333187745679167, + "grad_norm": 1.5481843948364258, + "learning_rate": 0.00016444645494062104, + "loss": 0.1264, + "step": 34190 + }, + { + "epoch": 0.5334747613402383, + "grad_norm": 1.8094528913497925, + "learning_rate": 0.00016443605582246626, + "loss": 0.2727, + "step": 34200 + }, + { + "epoch": 0.5336307481125601, + "grad_norm": 2.224538564682007, + "learning_rate": 0.00016442565670431148, + "loss": 0.3096, + "step": 34210 + }, + { + "epoch": 0.5337867348848817, + "grad_norm": 0.6375226974487305, + "learning_rate": 0.0001644152575861567, + "loss": 0.2251, + "step": 34220 + }, + { + "epoch": 0.5339427216572035, + "grad_norm": 3.727106809616089, + "learning_rate": 0.00016440485846800192, + "loss": 0.4374, + "step": 34230 + }, + { + "epoch": 0.5340987084295252, + "grad_norm": 0.13345426321029663, + "learning_rate": 0.00016439445934984714, + "loss": 0.2011, + "step": 34240 + }, + { + "epoch": 0.5342546952018469, + "grad_norm": 2.1658668518066406, + "learning_rate": 0.00016438406023169236, + "loss": 0.2457, + "step": 34250 + }, + { + "epoch": 0.5344106819741686, + "grad_norm": 19.238407135009766, + "learning_rate": 0.00016437366111353758, + "loss": 0.2756, + "step": 34260 + }, + { + "epoch": 0.5345666687464903, + "grad_norm": 1.0292778015136719, + "learning_rate": 0.0001643632619953828, + "loss": 0.1646, + "step": 34270 + }, + { + "epoch": 0.534722655518812, + "grad_norm": 0.9372987747192383, + "learning_rate": 0.00016435286287722802, + "loss": 0.2762, + "step": 34280 + }, + { + "epoch": 0.5348786422911337, + "grad_norm": 0.3918002247810364, + "learning_rate": 0.00016434246375907324, + "loss": 0.1406, + "step": 34290 + }, + { + "epoch": 0.5350346290634554, + "grad_norm": 1.3518732786178589, + "learning_rate": 0.00016433206464091846, + "loss": 0.2992, + "step": 34300 + }, + { + "epoch": 0.5351906158357771, + "grad_norm": 0.73117595911026, + "learning_rate": 0.00016432166552276368, + "loss": 0.1172, + "step": 34310 + }, + { + "epoch": 0.5353466026080989, + "grad_norm": 0.06655958294868469, + "learning_rate": 0.0001643112664046089, + "loss": 0.1332, + "step": 34320 + }, + { + "epoch": 0.5355025893804205, + "grad_norm": 3.2587168216705322, + "learning_rate": 0.00016430086728645412, + "loss": 0.2052, + "step": 34330 + }, + { + "epoch": 0.5356585761527423, + "grad_norm": 2.6668431758880615, + "learning_rate": 0.00016429046816829934, + "loss": 0.1295, + "step": 34340 + }, + { + "epoch": 0.5358145629250639, + "grad_norm": 1.316080927848816, + "learning_rate": 0.00016428006905014456, + "loss": 0.2072, + "step": 34350 + }, + { + "epoch": 0.5359705496973857, + "grad_norm": 5.144528388977051, + "learning_rate": 0.00016426966993198977, + "loss": 0.1934, + "step": 34360 + }, + { + "epoch": 0.5361265364697073, + "grad_norm": 1.3529599905014038, + "learning_rate": 0.000164259270813835, + "loss": 0.2077, + "step": 34370 + }, + { + "epoch": 0.5362825232420291, + "grad_norm": 0.9132925868034363, + "learning_rate": 0.00016424887169568021, + "loss": 0.2404, + "step": 34380 + }, + { + "epoch": 0.5364385100143508, + "grad_norm": 1.520033597946167, + "learning_rate": 0.00016423847257752543, + "loss": 0.1765, + "step": 34390 + }, + { + "epoch": 0.5365944967866725, + "grad_norm": 2.343975305557251, + "learning_rate": 0.00016422807345937065, + "loss": 0.1959, + "step": 34400 + }, + { + "epoch": 0.5367504835589942, + "grad_norm": 0.5855118632316589, + "learning_rate": 0.00016421767434121587, + "loss": 0.0848, + "step": 34410 + }, + { + "epoch": 0.5369064703313159, + "grad_norm": 1.9108648300170898, + "learning_rate": 0.0001642072752230611, + "loss": 0.2406, + "step": 34420 + }, + { + "epoch": 0.5370624571036376, + "grad_norm": 1.0323792695999146, + "learning_rate": 0.0001641968761049063, + "loss": 0.2298, + "step": 34430 + }, + { + "epoch": 0.5372184438759593, + "grad_norm": 0.17435620725154877, + "learning_rate": 0.00016418647698675153, + "loss": 0.1396, + "step": 34440 + }, + { + "epoch": 0.537374430648281, + "grad_norm": 0.3526577353477478, + "learning_rate": 0.00016417607786859675, + "loss": 0.172, + "step": 34450 + }, + { + "epoch": 0.5375304174206027, + "grad_norm": 2.73685884475708, + "learning_rate": 0.00016416567875044197, + "loss": 0.1167, + "step": 34460 + }, + { + "epoch": 0.5376864041929245, + "grad_norm": 2.765693426132202, + "learning_rate": 0.0001641552796322872, + "loss": 0.1356, + "step": 34470 + }, + { + "epoch": 0.5378423909652461, + "grad_norm": 0.47628021240234375, + "learning_rate": 0.0001641448805141324, + "loss": 0.3104, + "step": 34480 + }, + { + "epoch": 0.5379983777375679, + "grad_norm": 1.1307590007781982, + "learning_rate": 0.00016413448139597763, + "loss": 0.3247, + "step": 34490 + }, + { + "epoch": 0.5381543645098895, + "grad_norm": 0.07136381417512894, + "learning_rate": 0.00016412408227782285, + "loss": 0.319, + "step": 34500 + }, + { + "epoch": 0.5383103512822113, + "grad_norm": 1.7850221395492554, + "learning_rate": 0.00016411368315966807, + "loss": 0.2515, + "step": 34510 + }, + { + "epoch": 0.5384663380545329, + "grad_norm": 1.3870742321014404, + "learning_rate": 0.0001641032840415133, + "loss": 0.256, + "step": 34520 + }, + { + "epoch": 0.5386223248268547, + "grad_norm": 1.1329221725463867, + "learning_rate": 0.0001640928849233585, + "loss": 0.2255, + "step": 34530 + }, + { + "epoch": 0.5387783115991764, + "grad_norm": 0.23236137628555298, + "learning_rate": 0.00016408248580520373, + "loss": 0.4053, + "step": 34540 + }, + { + "epoch": 0.5389342983714981, + "grad_norm": 3.350924253463745, + "learning_rate": 0.00016407208668704895, + "loss": 0.2816, + "step": 34550 + }, + { + "epoch": 0.5390902851438198, + "grad_norm": 1.7459170818328857, + "learning_rate": 0.00016406168756889417, + "loss": 0.0834, + "step": 34560 + }, + { + "epoch": 0.5392462719161415, + "grad_norm": 0.89351487159729, + "learning_rate": 0.0001640512884507394, + "loss": 0.1146, + "step": 34570 + }, + { + "epoch": 0.5394022586884633, + "grad_norm": 0.49410831928253174, + "learning_rate": 0.0001640408893325846, + "loss": 0.2319, + "step": 34580 + }, + { + "epoch": 0.5395582454607849, + "grad_norm": 0.009971237741410732, + "learning_rate": 0.00016403049021442983, + "loss": 0.1674, + "step": 34590 + }, + { + "epoch": 0.5397142322331067, + "grad_norm": 1.420511245727539, + "learning_rate": 0.00016402009109627505, + "loss": 0.1469, + "step": 34600 + }, + { + "epoch": 0.5398702190054283, + "grad_norm": 2.5686769485473633, + "learning_rate": 0.00016400969197812027, + "loss": 0.2607, + "step": 34610 + }, + { + "epoch": 0.5400262057777501, + "grad_norm": 2.5808913707733154, + "learning_rate": 0.00016399929285996549, + "loss": 0.4412, + "step": 34620 + }, + { + "epoch": 0.5401821925500717, + "grad_norm": 0.2778591811656952, + "learning_rate": 0.0001639888937418107, + "loss": 0.1563, + "step": 34630 + }, + { + "epoch": 0.5403381793223935, + "grad_norm": 1.6448099613189697, + "learning_rate": 0.00016397849462365592, + "loss": 0.2484, + "step": 34640 + }, + { + "epoch": 0.5404941660947151, + "grad_norm": 0.003213417250663042, + "learning_rate": 0.00016396809550550114, + "loss": 0.1582, + "step": 34650 + }, + { + "epoch": 0.5406501528670369, + "grad_norm": 2.3433990478515625, + "learning_rate": 0.00016395769638734636, + "loss": 0.2774, + "step": 34660 + }, + { + "epoch": 0.5408061396393585, + "grad_norm": 0.09544426947832108, + "learning_rate": 0.00016394729726919158, + "loss": 0.1242, + "step": 34670 + }, + { + "epoch": 0.5409621264116803, + "grad_norm": 1.8934235572814941, + "learning_rate": 0.0001639368981510368, + "loss": 0.1636, + "step": 34680 + }, + { + "epoch": 0.541118113184002, + "grad_norm": 2.8528554439544678, + "learning_rate": 0.00016392649903288202, + "loss": 0.2524, + "step": 34690 + }, + { + "epoch": 0.5412740999563237, + "grad_norm": 2.837836980819702, + "learning_rate": 0.00016391609991472722, + "loss": 0.1808, + "step": 34700 + }, + { + "epoch": 0.5414300867286455, + "grad_norm": 4.349740028381348, + "learning_rate": 0.00016390570079657246, + "loss": 0.12, + "step": 34710 + }, + { + "epoch": 0.5415860735009671, + "grad_norm": 0.6011945605278015, + "learning_rate": 0.00016389530167841765, + "loss": 0.2865, + "step": 34720 + }, + { + "epoch": 0.5417420602732889, + "grad_norm": 1.2079488039016724, + "learning_rate": 0.0001638849025602629, + "loss": 0.125, + "step": 34730 + }, + { + "epoch": 0.5418980470456105, + "grad_norm": 1.9549164772033691, + "learning_rate": 0.0001638745034421081, + "loss": 0.3285, + "step": 34740 + }, + { + "epoch": 0.5420540338179323, + "grad_norm": 0.6969407200813293, + "learning_rate": 0.00016386410432395334, + "loss": 0.3267, + "step": 34750 + }, + { + "epoch": 0.5422100205902539, + "grad_norm": 1.0561965703964233, + "learning_rate": 0.00016385370520579853, + "loss": 0.1463, + "step": 34760 + }, + { + "epoch": 0.5423660073625757, + "grad_norm": 1.8208931684494019, + "learning_rate": 0.00016384330608764378, + "loss": 0.1713, + "step": 34770 + }, + { + "epoch": 0.5425219941348973, + "grad_norm": 0.6349910497665405, + "learning_rate": 0.000163832906969489, + "loss": 0.2244, + "step": 34780 + }, + { + "epoch": 0.5426779809072191, + "grad_norm": 1.1976515054702759, + "learning_rate": 0.00016382250785133422, + "loss": 0.3407, + "step": 34790 + }, + { + "epoch": 0.5428339676795407, + "grad_norm": 0.5970319509506226, + "learning_rate": 0.00016381210873317944, + "loss": 0.1419, + "step": 34800 + }, + { + "epoch": 0.5429899544518625, + "grad_norm": 0.14623159170150757, + "learning_rate": 0.00016380170961502466, + "loss": 0.14, + "step": 34810 + }, + { + "epoch": 0.5431459412241841, + "grad_norm": 1.4332351684570312, + "learning_rate": 0.00016379131049686988, + "loss": 0.1093, + "step": 34820 + }, + { + "epoch": 0.5433019279965059, + "grad_norm": 1.1443063020706177, + "learning_rate": 0.0001637809113787151, + "loss": 0.3676, + "step": 34830 + }, + { + "epoch": 0.5434579147688277, + "grad_norm": 1.3897461891174316, + "learning_rate": 0.00016377051226056032, + "loss": 0.3713, + "step": 34840 + }, + { + "epoch": 0.5436139015411493, + "grad_norm": 1.7773199081420898, + "learning_rate": 0.00016376011314240554, + "loss": 0.2249, + "step": 34850 + }, + { + "epoch": 0.5437698883134711, + "grad_norm": 0.8524389266967773, + "learning_rate": 0.00016374971402425076, + "loss": 0.2158, + "step": 34860 + }, + { + "epoch": 0.5439258750857927, + "grad_norm": 0.5572389960289001, + "learning_rate": 0.00016373931490609598, + "loss": 0.215, + "step": 34870 + }, + { + "epoch": 0.5440818618581145, + "grad_norm": 3.0238709449768066, + "learning_rate": 0.0001637289157879412, + "loss": 0.3765, + "step": 34880 + }, + { + "epoch": 0.5442378486304361, + "grad_norm": 1.1655421257019043, + "learning_rate": 0.00016371851666978642, + "loss": 0.2491, + "step": 34890 + }, + { + "epoch": 0.5443938354027579, + "grad_norm": 0.49492090940475464, + "learning_rate": 0.00016370811755163164, + "loss": 0.1287, + "step": 34900 + }, + { + "epoch": 0.5445498221750795, + "grad_norm": 0.2732921540737152, + "learning_rate": 0.00016369771843347685, + "loss": 0.2154, + "step": 34910 + }, + { + "epoch": 0.5447058089474013, + "grad_norm": 2.5807769298553467, + "learning_rate": 0.00016368731931532207, + "loss": 0.2032, + "step": 34920 + }, + { + "epoch": 0.5448617957197229, + "grad_norm": 0.5044315457344055, + "learning_rate": 0.0001636769201971673, + "loss": 0.2595, + "step": 34930 + }, + { + "epoch": 0.5450177824920447, + "grad_norm": 2.1248972415924072, + "learning_rate": 0.00016366652107901251, + "loss": 0.5651, + "step": 34940 + }, + { + "epoch": 0.5451737692643663, + "grad_norm": 0.8391468524932861, + "learning_rate": 0.00016365612196085773, + "loss": 0.3028, + "step": 34950 + }, + { + "epoch": 0.5453297560366881, + "grad_norm": 0.36081477999687195, + "learning_rate": 0.00016364572284270295, + "loss": 0.1073, + "step": 34960 + }, + { + "epoch": 0.5454857428090097, + "grad_norm": 3.175804376602173, + "learning_rate": 0.00016363532372454817, + "loss": 0.1889, + "step": 34970 + }, + { + "epoch": 0.5456417295813315, + "grad_norm": 0.20647937059402466, + "learning_rate": 0.0001636249246063934, + "loss": 0.2416, + "step": 34980 + }, + { + "epoch": 0.5457977163536533, + "grad_norm": 2.3525943756103516, + "learning_rate": 0.0001636145254882386, + "loss": 0.3179, + "step": 34990 + }, + { + "epoch": 0.5459537031259749, + "grad_norm": 0.2641216814517975, + "learning_rate": 0.00016360412637008383, + "loss": 0.2157, + "step": 35000 + }, + { + "epoch": 0.5461096898982967, + "grad_norm": 0.09906073659658432, + "learning_rate": 0.00016359372725192905, + "loss": 0.2576, + "step": 35010 + }, + { + "epoch": 0.5462656766706183, + "grad_norm": 0.13971920311450958, + "learning_rate": 0.00016358332813377427, + "loss": 0.1298, + "step": 35020 + }, + { + "epoch": 0.5464216634429401, + "grad_norm": 1.4535638093948364, + "learning_rate": 0.0001635729290156195, + "loss": 0.2357, + "step": 35030 + }, + { + "epoch": 0.5465776502152617, + "grad_norm": 0.9107828736305237, + "learning_rate": 0.0001635625298974647, + "loss": 0.3049, + "step": 35040 + }, + { + "epoch": 0.5467336369875835, + "grad_norm": 0.0377386212348938, + "learning_rate": 0.00016355213077930993, + "loss": 0.2066, + "step": 35050 + }, + { + "epoch": 0.5468896237599051, + "grad_norm": 2.7800869941711426, + "learning_rate": 0.00016354173166115515, + "loss": 0.2974, + "step": 35060 + }, + { + "epoch": 0.5470456105322269, + "grad_norm": 1.6247998476028442, + "learning_rate": 0.00016353133254300037, + "loss": 0.3822, + "step": 35070 + }, + { + "epoch": 0.5472015973045485, + "grad_norm": 1.6479015350341797, + "learning_rate": 0.0001635209334248456, + "loss": 0.3401, + "step": 35080 + }, + { + "epoch": 0.5473575840768703, + "grad_norm": 0.1879737675189972, + "learning_rate": 0.0001635105343066908, + "loss": 0.1549, + "step": 35090 + }, + { + "epoch": 0.5475135708491919, + "grad_norm": 0.39355361461639404, + "learning_rate": 0.00016350013518853603, + "loss": 0.2032, + "step": 35100 + }, + { + "epoch": 0.5476695576215137, + "grad_norm": 2.9976983070373535, + "learning_rate": 0.00016348973607038125, + "loss": 0.3867, + "step": 35110 + }, + { + "epoch": 0.5478255443938354, + "grad_norm": 0.8242707252502441, + "learning_rate": 0.00016347933695222647, + "loss": 0.3601, + "step": 35120 + }, + { + "epoch": 0.5479815311661571, + "grad_norm": 0.874955415725708, + "learning_rate": 0.0001634689378340717, + "loss": 0.2709, + "step": 35130 + }, + { + "epoch": 0.5481375179384789, + "grad_norm": 0.05980971083045006, + "learning_rate": 0.0001634585387159169, + "loss": 0.2646, + "step": 35140 + }, + { + "epoch": 0.5482935047108005, + "grad_norm": 1.8437328338623047, + "learning_rate": 0.0001634481395977621, + "loss": 0.1965, + "step": 35150 + }, + { + "epoch": 0.5484494914831223, + "grad_norm": 0.4142405688762665, + "learning_rate": 0.00016343774047960735, + "loss": 0.1756, + "step": 35160 + }, + { + "epoch": 0.5486054782554439, + "grad_norm": 4.836324214935303, + "learning_rate": 0.00016342734136145254, + "loss": 0.2534, + "step": 35170 + }, + { + "epoch": 0.5487614650277657, + "grad_norm": 1.4933065176010132, + "learning_rate": 0.00016341694224329779, + "loss": 0.2355, + "step": 35180 + }, + { + "epoch": 0.5489174518000873, + "grad_norm": 1.9468894004821777, + "learning_rate": 0.00016340654312514298, + "loss": 0.3113, + "step": 35190 + }, + { + "epoch": 0.5490734385724091, + "grad_norm": 1.012710452079773, + "learning_rate": 0.00016339614400698822, + "loss": 0.2635, + "step": 35200 + }, + { + "epoch": 0.5492294253447307, + "grad_norm": 1.5551140308380127, + "learning_rate": 0.00016338574488883342, + "loss": 0.2951, + "step": 35210 + }, + { + "epoch": 0.5493854121170525, + "grad_norm": 1.8613696098327637, + "learning_rate": 0.00016337534577067866, + "loss": 0.2832, + "step": 35220 + }, + { + "epoch": 0.5495413988893741, + "grad_norm": 1.6748839616775513, + "learning_rate": 0.00016336494665252386, + "loss": 0.2266, + "step": 35230 + }, + { + "epoch": 0.5496973856616959, + "grad_norm": 0.4094032347202301, + "learning_rate": 0.0001633545475343691, + "loss": 0.3497, + "step": 35240 + }, + { + "epoch": 0.5498533724340176, + "grad_norm": 1.6414631605148315, + "learning_rate": 0.0001633441484162143, + "loss": 0.1763, + "step": 35250 + }, + { + "epoch": 0.5500093592063393, + "grad_norm": 1.6380645036697388, + "learning_rate": 0.00016333374929805954, + "loss": 0.1355, + "step": 35260 + }, + { + "epoch": 0.550165345978661, + "grad_norm": 0.7325630187988281, + "learning_rate": 0.00016332335017990473, + "loss": 0.3105, + "step": 35270 + }, + { + "epoch": 0.5503213327509827, + "grad_norm": 1.6015644073486328, + "learning_rate": 0.00016331295106174998, + "loss": 0.2311, + "step": 35280 + }, + { + "epoch": 0.5504773195233045, + "grad_norm": 0.9224210381507874, + "learning_rate": 0.00016330255194359517, + "loss": 0.2746, + "step": 35290 + }, + { + "epoch": 0.5506333062956261, + "grad_norm": 1.30025315284729, + "learning_rate": 0.00016329215282544042, + "loss": 0.2066, + "step": 35300 + }, + { + "epoch": 0.5507892930679479, + "grad_norm": 0.481125146150589, + "learning_rate": 0.0001632817537072856, + "loss": 0.1778, + "step": 35310 + }, + { + "epoch": 0.5509452798402695, + "grad_norm": 0.2709486782550812, + "learning_rate": 0.00016327135458913086, + "loss": 0.1131, + "step": 35320 + }, + { + "epoch": 0.5511012666125913, + "grad_norm": 0.6248563528060913, + "learning_rate": 0.00016326095547097605, + "loss": 0.4356, + "step": 35330 + }, + { + "epoch": 0.5512572533849129, + "grad_norm": 3.0947647094726562, + "learning_rate": 0.0001632505563528213, + "loss": 0.3072, + "step": 35340 + }, + { + "epoch": 0.5514132401572347, + "grad_norm": 1.009535312652588, + "learning_rate": 0.0001632401572346665, + "loss": 0.1808, + "step": 35350 + }, + { + "epoch": 0.5515692269295563, + "grad_norm": 0.042604975402355194, + "learning_rate": 0.00016322975811651174, + "loss": 0.1198, + "step": 35360 + }, + { + "epoch": 0.5517252137018781, + "grad_norm": 1.8652396202087402, + "learning_rate": 0.00016321935899835693, + "loss": 0.2176, + "step": 35370 + }, + { + "epoch": 0.5518812004741998, + "grad_norm": 4.6097187995910645, + "learning_rate": 0.00016320895988020218, + "loss": 0.2132, + "step": 35380 + }, + { + "epoch": 0.5520371872465215, + "grad_norm": 1.3615522384643555, + "learning_rate": 0.00016319856076204737, + "loss": 0.2751, + "step": 35390 + }, + { + "epoch": 0.5521931740188432, + "grad_norm": 0.8844773769378662, + "learning_rate": 0.00016318816164389262, + "loss": 0.22, + "step": 35400 + }, + { + "epoch": 0.5523491607911649, + "grad_norm": 0.5413331985473633, + "learning_rate": 0.0001631777625257378, + "loss": 0.2159, + "step": 35410 + }, + { + "epoch": 0.5525051475634866, + "grad_norm": 1.5456678867340088, + "learning_rate": 0.00016316736340758306, + "loss": 0.2244, + "step": 35420 + }, + { + "epoch": 0.5526611343358083, + "grad_norm": 2.1405861377716064, + "learning_rate": 0.00016315696428942825, + "loss": 0.446, + "step": 35430 + }, + { + "epoch": 0.5528171211081301, + "grad_norm": 1.4269858598709106, + "learning_rate": 0.0001631465651712735, + "loss": 0.3599, + "step": 35440 + }, + { + "epoch": 0.5529731078804517, + "grad_norm": 1.8682516813278198, + "learning_rate": 0.0001631361660531187, + "loss": 0.115, + "step": 35450 + }, + { + "epoch": 0.5531290946527735, + "grad_norm": 1.0175774097442627, + "learning_rate": 0.00016312576693496394, + "loss": 0.4003, + "step": 35460 + }, + { + "epoch": 0.5532850814250951, + "grad_norm": 2.4369170665740967, + "learning_rate": 0.00016311536781680913, + "loss": 0.2489, + "step": 35470 + }, + { + "epoch": 0.5534410681974169, + "grad_norm": 1.886022686958313, + "learning_rate": 0.00016310496869865437, + "loss": 0.4201, + "step": 35480 + }, + { + "epoch": 0.5535970549697385, + "grad_norm": 1.5811959505081177, + "learning_rate": 0.00016309456958049957, + "loss": 0.4286, + "step": 35490 + }, + { + "epoch": 0.5537530417420603, + "grad_norm": 1.971110224723816, + "learning_rate": 0.00016308417046234481, + "loss": 0.3599, + "step": 35500 + }, + { + "epoch": 0.553909028514382, + "grad_norm": 0.0335380844771862, + "learning_rate": 0.00016307377134419, + "loss": 0.3357, + "step": 35510 + }, + { + "epoch": 0.5540650152867037, + "grad_norm": 1.8519577980041504, + "learning_rate": 0.00016306337222603525, + "loss": 0.3126, + "step": 35520 + }, + { + "epoch": 0.5542210020590254, + "grad_norm": 2.1563379764556885, + "learning_rate": 0.00016305297310788045, + "loss": 0.1123, + "step": 35530 + }, + { + "epoch": 0.5543769888313471, + "grad_norm": 1.64332914352417, + "learning_rate": 0.0001630425739897257, + "loss": 0.2853, + "step": 35540 + }, + { + "epoch": 0.5545329756036688, + "grad_norm": 0.061150554567575455, + "learning_rate": 0.00016303217487157088, + "loss": 0.1943, + "step": 35550 + }, + { + "epoch": 0.5546889623759905, + "grad_norm": 1.2701060771942139, + "learning_rate": 0.00016302177575341613, + "loss": 0.2304, + "step": 35560 + }, + { + "epoch": 0.5548449491483122, + "grad_norm": 2.424860715866089, + "learning_rate": 0.00016301137663526132, + "loss": 0.2128, + "step": 35570 + }, + { + "epoch": 0.5550009359206339, + "grad_norm": 0.6803575158119202, + "learning_rate": 0.00016300097751710657, + "loss": 0.1145, + "step": 35580 + }, + { + "epoch": 0.5551569226929557, + "grad_norm": 1.2855092287063599, + "learning_rate": 0.00016299057839895176, + "loss": 0.2448, + "step": 35590 + }, + { + "epoch": 0.5553129094652773, + "grad_norm": 2.6340911388397217, + "learning_rate": 0.00016298017928079698, + "loss": 0.1856, + "step": 35600 + }, + { + "epoch": 0.5554688962375991, + "grad_norm": 1.776382327079773, + "learning_rate": 0.0001629697801626422, + "loss": 0.2422, + "step": 35610 + }, + { + "epoch": 0.5556248830099207, + "grad_norm": 3.0746347904205322, + "learning_rate": 0.00016295938104448742, + "loss": 0.1578, + "step": 35620 + }, + { + "epoch": 0.5557808697822425, + "grad_norm": 1.4887659549713135, + "learning_rate": 0.00016294898192633267, + "loss": 0.2385, + "step": 35630 + }, + { + "epoch": 0.5559368565545642, + "grad_norm": 0.5371220707893372, + "learning_rate": 0.00016293858280817786, + "loss": 0.2361, + "step": 35640 + }, + { + "epoch": 0.5560928433268859, + "grad_norm": 0.18604904413223267, + "learning_rate": 0.0001629281836900231, + "loss": 0.0828, + "step": 35650 + }, + { + "epoch": 0.5562488300992076, + "grad_norm": 2.267854690551758, + "learning_rate": 0.0001629177845718683, + "loss": 0.413, + "step": 35660 + }, + { + "epoch": 0.5564048168715293, + "grad_norm": 1.9755452871322632, + "learning_rate": 0.00016290738545371355, + "loss": 0.146, + "step": 35670 + }, + { + "epoch": 0.556560803643851, + "grad_norm": 2.430293321609497, + "learning_rate": 0.00016289698633555874, + "loss": 0.2677, + "step": 35680 + }, + { + "epoch": 0.5567167904161727, + "grad_norm": 0.6319543719291687, + "learning_rate": 0.000162886587217404, + "loss": 0.1481, + "step": 35690 + }, + { + "epoch": 0.5568727771884944, + "grad_norm": 0.17439277470111847, + "learning_rate": 0.00016287618809924918, + "loss": 0.1641, + "step": 35700 + }, + { + "epoch": 0.5570287639608161, + "grad_norm": 0.350175142288208, + "learning_rate": 0.00016286578898109443, + "loss": 0.393, + "step": 35710 + }, + { + "epoch": 0.5571847507331378, + "grad_norm": 2.641941547393799, + "learning_rate": 0.00016285538986293962, + "loss": 0.1907, + "step": 35720 + }, + { + "epoch": 0.5573407375054595, + "grad_norm": 0.1910303384065628, + "learning_rate": 0.00016284499074478487, + "loss": 0.2248, + "step": 35730 + }, + { + "epoch": 0.5574967242777813, + "grad_norm": 1.0180896520614624, + "learning_rate": 0.00016283459162663006, + "loss": 0.1924, + "step": 35740 + }, + { + "epoch": 0.557652711050103, + "grad_norm": 1.2047260999679565, + "learning_rate": 0.0001628241925084753, + "loss": 0.2031, + "step": 35750 + }, + { + "epoch": 0.5578086978224247, + "grad_norm": 1.8182405233383179, + "learning_rate": 0.0001628137933903205, + "loss": 0.1573, + "step": 35760 + }, + { + "epoch": 0.5579646845947464, + "grad_norm": 0.5485963821411133, + "learning_rate": 0.00016280339427216574, + "loss": 0.152, + "step": 35770 + }, + { + "epoch": 0.5581206713670681, + "grad_norm": 0.1674145758152008, + "learning_rate": 0.00016279299515401094, + "loss": 0.2388, + "step": 35780 + }, + { + "epoch": 0.5582766581393898, + "grad_norm": 0.36187657713890076, + "learning_rate": 0.00016278259603585618, + "loss": 0.1618, + "step": 35790 + }, + { + "epoch": 0.5584326449117115, + "grad_norm": 0.013768521137535572, + "learning_rate": 0.00016277219691770138, + "loss": 0.3666, + "step": 35800 + }, + { + "epoch": 0.5585886316840332, + "grad_norm": 2.8602802753448486, + "learning_rate": 0.00016276179779954662, + "loss": 0.2065, + "step": 35810 + }, + { + "epoch": 0.5587446184563549, + "grad_norm": 1.7925455570220947, + "learning_rate": 0.00016275139868139182, + "loss": 0.1729, + "step": 35820 + }, + { + "epoch": 0.5589006052286766, + "grad_norm": 2.910456418991089, + "learning_rate": 0.00016274099956323706, + "loss": 0.3734, + "step": 35830 + }, + { + "epoch": 0.5590565920009983, + "grad_norm": 0.8378308415412903, + "learning_rate": 0.00016273060044508225, + "loss": 0.2255, + "step": 35840 + }, + { + "epoch": 0.55921257877332, + "grad_norm": 0.409534752368927, + "learning_rate": 0.0001627202013269275, + "loss": 0.2046, + "step": 35850 + }, + { + "epoch": 0.5593685655456417, + "grad_norm": 0.49498099088668823, + "learning_rate": 0.0001627098022087727, + "loss": 0.1058, + "step": 35860 + }, + { + "epoch": 0.5595245523179634, + "grad_norm": 1.1761783361434937, + "learning_rate": 0.00016269940309061794, + "loss": 0.1566, + "step": 35870 + }, + { + "epoch": 0.5596805390902851, + "grad_norm": 3.8204751014709473, + "learning_rate": 0.00016268900397246313, + "loss": 0.2646, + "step": 35880 + }, + { + "epoch": 0.5598365258626069, + "grad_norm": 0.9882522225379944, + "learning_rate": 0.00016267860485430838, + "loss": 0.1756, + "step": 35890 + }, + { + "epoch": 0.5599925126349286, + "grad_norm": 1.1832259893417358, + "learning_rate": 0.00016266820573615357, + "loss": 0.1385, + "step": 35900 + }, + { + "epoch": 0.5601484994072503, + "grad_norm": 0.7638296484947205, + "learning_rate": 0.00016265780661799882, + "loss": 0.5679, + "step": 35910 + }, + { + "epoch": 0.560304486179572, + "grad_norm": 0.8551504611968994, + "learning_rate": 0.000162647407499844, + "loss": 0.2981, + "step": 35920 + }, + { + "epoch": 0.5604604729518937, + "grad_norm": 5.499948501586914, + "learning_rate": 0.00016263700838168926, + "loss": 0.1429, + "step": 35930 + }, + { + "epoch": 0.5606164597242154, + "grad_norm": 3.0494496822357178, + "learning_rate": 0.00016262660926353445, + "loss": 0.1322, + "step": 35940 + }, + { + "epoch": 0.5607724464965371, + "grad_norm": 0.2482384592294693, + "learning_rate": 0.0001626162101453797, + "loss": 0.2473, + "step": 35950 + }, + { + "epoch": 0.5609284332688588, + "grad_norm": 0.009750776924192905, + "learning_rate": 0.0001626058110272249, + "loss": 0.0429, + "step": 35960 + }, + { + "epoch": 0.5610844200411805, + "grad_norm": 0.5742604732513428, + "learning_rate": 0.00016259541190907014, + "loss": 0.4565, + "step": 35970 + }, + { + "epoch": 0.5612404068135022, + "grad_norm": 0.7810243368148804, + "learning_rate": 0.00016258501279091533, + "loss": 0.2107, + "step": 35980 + }, + { + "epoch": 0.5613963935858239, + "grad_norm": 2.785747528076172, + "learning_rate": 0.00016257461367276058, + "loss": 0.5087, + "step": 35990 + }, + { + "epoch": 0.5615523803581456, + "grad_norm": 0.8806902766227722, + "learning_rate": 0.00016256421455460577, + "loss": 0.2937, + "step": 36000 + }, + { + "epoch": 0.5617083671304673, + "grad_norm": 1.857373833656311, + "learning_rate": 0.00016255381543645102, + "loss": 0.5042, + "step": 36010 + }, + { + "epoch": 0.561864353902789, + "grad_norm": 0.5333901047706604, + "learning_rate": 0.0001625434163182962, + "loss": 0.1678, + "step": 36020 + }, + { + "epoch": 0.5620203406751108, + "grad_norm": 11.023160934448242, + "learning_rate": 0.00016253301720014145, + "loss": 0.2721, + "step": 36030 + }, + { + "epoch": 0.5621763274474325, + "grad_norm": 2.6214029788970947, + "learning_rate": 0.00016252261808198665, + "loss": 0.1675, + "step": 36040 + }, + { + "epoch": 0.5623323142197542, + "grad_norm": 0.25980064272880554, + "learning_rate": 0.0001625122189638319, + "loss": 0.1832, + "step": 36050 + }, + { + "epoch": 0.5624883009920759, + "grad_norm": 1.3559473752975464, + "learning_rate": 0.0001625018198456771, + "loss": 0.121, + "step": 36060 + }, + { + "epoch": 0.5626442877643976, + "grad_norm": 0.3558153212070465, + "learning_rate": 0.0001624914207275223, + "loss": 0.2797, + "step": 36070 + }, + { + "epoch": 0.5628002745367193, + "grad_norm": 0.6314427256584167, + "learning_rate": 0.00016248102160936753, + "loss": 0.3455, + "step": 36080 + }, + { + "epoch": 0.562956261309041, + "grad_norm": 0.16120101511478424, + "learning_rate": 0.00016247062249121275, + "loss": 0.3669, + "step": 36090 + }, + { + "epoch": 0.5631122480813627, + "grad_norm": 1.041572093963623, + "learning_rate": 0.00016246022337305797, + "loss": 0.2744, + "step": 36100 + }, + { + "epoch": 0.5632682348536844, + "grad_norm": 0.2645890414714813, + "learning_rate": 0.00016244982425490318, + "loss": 0.1487, + "step": 36110 + }, + { + "epoch": 0.5634242216260061, + "grad_norm": 5.698698043823242, + "learning_rate": 0.0001624394251367484, + "loss": 0.173, + "step": 36120 + }, + { + "epoch": 0.5635802083983278, + "grad_norm": 0.3299804627895355, + "learning_rate": 0.00016242902601859362, + "loss": 0.1965, + "step": 36130 + }, + { + "epoch": 0.5637361951706495, + "grad_norm": 1.2476481199264526, + "learning_rate": 0.00016241862690043884, + "loss": 0.4097, + "step": 36140 + }, + { + "epoch": 0.5638921819429712, + "grad_norm": 0.30926263332366943, + "learning_rate": 0.00016240822778228406, + "loss": 0.1416, + "step": 36150 + }, + { + "epoch": 0.564048168715293, + "grad_norm": 0.05872740224003792, + "learning_rate": 0.00016239782866412928, + "loss": 0.1449, + "step": 36160 + }, + { + "epoch": 0.5642041554876146, + "grad_norm": 0.2292211800813675, + "learning_rate": 0.0001623874295459745, + "loss": 0.2557, + "step": 36170 + }, + { + "epoch": 0.5643601422599364, + "grad_norm": 1.7822531461715698, + "learning_rate": 0.00016237703042781972, + "loss": 0.2953, + "step": 36180 + }, + { + "epoch": 0.5645161290322581, + "grad_norm": 2.4908461570739746, + "learning_rate": 0.00016236663130966494, + "loss": 0.3888, + "step": 36190 + }, + { + "epoch": 0.5646721158045798, + "grad_norm": 0.08363594114780426, + "learning_rate": 0.00016235623219151016, + "loss": 0.2293, + "step": 36200 + }, + { + "epoch": 0.5648281025769015, + "grad_norm": 1.4830002784729004, + "learning_rate": 0.00016234583307335538, + "loss": 0.2348, + "step": 36210 + }, + { + "epoch": 0.5649840893492232, + "grad_norm": 1.8443071842193604, + "learning_rate": 0.0001623354339552006, + "loss": 0.3392, + "step": 36220 + }, + { + "epoch": 0.5651400761215449, + "grad_norm": 2.8050875663757324, + "learning_rate": 0.00016232503483704582, + "loss": 0.2426, + "step": 36230 + }, + { + "epoch": 0.5652960628938666, + "grad_norm": 3.6627259254455566, + "learning_rate": 0.00016231463571889104, + "loss": 0.2118, + "step": 36240 + }, + { + "epoch": 0.5654520496661883, + "grad_norm": 0.3443094789981842, + "learning_rate": 0.00016230423660073626, + "loss": 0.2418, + "step": 36250 + }, + { + "epoch": 0.56560803643851, + "grad_norm": 1.7411112785339355, + "learning_rate": 0.00016229383748258148, + "loss": 0.1628, + "step": 36260 + }, + { + "epoch": 0.5657640232108317, + "grad_norm": 1.1321816444396973, + "learning_rate": 0.0001622834383644267, + "loss": 0.2732, + "step": 36270 + }, + { + "epoch": 0.5659200099831534, + "grad_norm": 2.9686615467071533, + "learning_rate": 0.00016227303924627192, + "loss": 0.3036, + "step": 36280 + }, + { + "epoch": 0.5660759967554752, + "grad_norm": 5.03530740737915, + "learning_rate": 0.00016226264012811714, + "loss": 0.2143, + "step": 36290 + }, + { + "epoch": 0.5662319835277968, + "grad_norm": 2.3943281173706055, + "learning_rate": 0.00016225224100996236, + "loss": 0.2629, + "step": 36300 + }, + { + "epoch": 0.5663879703001186, + "grad_norm": 1.812828779220581, + "learning_rate": 0.00016224184189180758, + "loss": 0.3892, + "step": 36310 + }, + { + "epoch": 0.5665439570724402, + "grad_norm": 1.4708483219146729, + "learning_rate": 0.0001622314427736528, + "loss": 0.2118, + "step": 36320 + }, + { + "epoch": 0.566699943844762, + "grad_norm": 8.99713134765625, + "learning_rate": 0.00016222104365549802, + "loss": 0.3185, + "step": 36330 + }, + { + "epoch": 0.5668559306170837, + "grad_norm": 1.7472341060638428, + "learning_rate": 0.00016221064453734324, + "loss": 0.1659, + "step": 36340 + }, + { + "epoch": 0.5670119173894054, + "grad_norm": 0.8059778213500977, + "learning_rate": 0.00016220024541918846, + "loss": 0.0621, + "step": 36350 + }, + { + "epoch": 0.5671679041617271, + "grad_norm": 4.0501861572265625, + "learning_rate": 0.00016218984630103368, + "loss": 0.1399, + "step": 36360 + }, + { + "epoch": 0.5673238909340488, + "grad_norm": 0.014400321058928967, + "learning_rate": 0.0001621794471828789, + "loss": 0.2372, + "step": 36370 + }, + { + "epoch": 0.5674798777063705, + "grad_norm": 0.4061933755874634, + "learning_rate": 0.00016216904806472412, + "loss": 0.3854, + "step": 36380 + }, + { + "epoch": 0.5676358644786922, + "grad_norm": 0.4940955340862274, + "learning_rate": 0.00016215864894656933, + "loss": 0.1689, + "step": 36390 + }, + { + "epoch": 0.567791851251014, + "grad_norm": 0.3220517635345459, + "learning_rate": 0.00016214824982841455, + "loss": 0.1912, + "step": 36400 + }, + { + "epoch": 0.5679478380233356, + "grad_norm": 2.0848920345306396, + "learning_rate": 0.00016213785071025977, + "loss": 0.1639, + "step": 36410 + }, + { + "epoch": 0.5681038247956574, + "grad_norm": 0.04129406809806824, + "learning_rate": 0.000162127451592105, + "loss": 0.1378, + "step": 36420 + }, + { + "epoch": 0.568259811567979, + "grad_norm": 1.9343401193618774, + "learning_rate": 0.0001621170524739502, + "loss": 0.3759, + "step": 36430 + }, + { + "epoch": 0.5684157983403008, + "grad_norm": 1.0625615119934082, + "learning_rate": 0.00016210665335579543, + "loss": 0.2074, + "step": 36440 + }, + { + "epoch": 0.5685717851126224, + "grad_norm": 0.14784491062164307, + "learning_rate": 0.00016209625423764065, + "loss": 0.0981, + "step": 36450 + }, + { + "epoch": 0.5687277718849442, + "grad_norm": 1.8117821216583252, + "learning_rate": 0.00016208585511948587, + "loss": 0.2755, + "step": 36460 + }, + { + "epoch": 0.5688837586572658, + "grad_norm": 0.061010006815195084, + "learning_rate": 0.0001620754560013311, + "loss": 0.1815, + "step": 36470 + }, + { + "epoch": 0.5690397454295876, + "grad_norm": 0.44453224539756775, + "learning_rate": 0.0001620650568831763, + "loss": 0.0891, + "step": 36480 + }, + { + "epoch": 0.5691957322019093, + "grad_norm": 0.6130645871162415, + "learning_rate": 0.00016205465776502153, + "loss": 0.2778, + "step": 36490 + }, + { + "epoch": 0.569351718974231, + "grad_norm": 5.408796787261963, + "learning_rate": 0.00016204425864686678, + "loss": 0.2301, + "step": 36500 + }, + { + "epoch": 0.5695077057465527, + "grad_norm": 1.2095412015914917, + "learning_rate": 0.00016203385952871197, + "loss": 0.1378, + "step": 36510 + }, + { + "epoch": 0.5696636925188744, + "grad_norm": 0.46873530745506287, + "learning_rate": 0.0001620234604105572, + "loss": 0.1431, + "step": 36520 + }, + { + "epoch": 0.5698196792911961, + "grad_norm": 1.028438925743103, + "learning_rate": 0.0001620130612924024, + "loss": 0.1114, + "step": 36530 + }, + { + "epoch": 0.5699756660635178, + "grad_norm": 0.5952587127685547, + "learning_rate": 0.00016200266217424763, + "loss": 0.2156, + "step": 36540 + }, + { + "epoch": 0.5701316528358396, + "grad_norm": 0.5134285688400269, + "learning_rate": 0.00016199226305609285, + "loss": 0.2712, + "step": 36550 + }, + { + "epoch": 0.5702876396081612, + "grad_norm": 0.03440163657069206, + "learning_rate": 0.00016198186393793807, + "loss": 0.312, + "step": 36560 + }, + { + "epoch": 0.570443626380483, + "grad_norm": 0.10363951325416565, + "learning_rate": 0.0001619714648197833, + "loss": 0.2399, + "step": 36570 + }, + { + "epoch": 0.5705996131528046, + "grad_norm": 1.3341947793960571, + "learning_rate": 0.0001619610657016285, + "loss": 0.1397, + "step": 36580 + }, + { + "epoch": 0.5707555999251264, + "grad_norm": 0.0668073296546936, + "learning_rate": 0.00016195066658347373, + "loss": 0.1091, + "step": 36590 + }, + { + "epoch": 0.570911586697448, + "grad_norm": 5.0486674308776855, + "learning_rate": 0.00016194026746531895, + "loss": 0.4893, + "step": 36600 + }, + { + "epoch": 0.5710675734697698, + "grad_norm": 1.2655692100524902, + "learning_rate": 0.00016192986834716417, + "loss": 0.3344, + "step": 36610 + }, + { + "epoch": 0.5712235602420914, + "grad_norm": 0.1458190381526947, + "learning_rate": 0.00016191946922900939, + "loss": 0.247, + "step": 36620 + }, + { + "epoch": 0.5713795470144132, + "grad_norm": 2.5789072513580322, + "learning_rate": 0.0001619090701108546, + "loss": 0.1546, + "step": 36630 + }, + { + "epoch": 0.5715355337867349, + "grad_norm": 1.191821575164795, + "learning_rate": 0.00016189867099269983, + "loss": 0.2215, + "step": 36640 + }, + { + "epoch": 0.5716915205590566, + "grad_norm": 8.163617134094238, + "learning_rate": 0.00016188827187454505, + "loss": 0.1696, + "step": 36650 + }, + { + "epoch": 0.5718475073313783, + "grad_norm": 3.162277936935425, + "learning_rate": 0.00016187787275639027, + "loss": 0.2906, + "step": 36660 + }, + { + "epoch": 0.5720034941037, + "grad_norm": 2.2368602752685547, + "learning_rate": 0.00016186747363823548, + "loss": 0.3806, + "step": 36670 + }, + { + "epoch": 0.5721594808760218, + "grad_norm": 2.0193803310394287, + "learning_rate": 0.0001618570745200807, + "loss": 0.1964, + "step": 36680 + }, + { + "epoch": 0.5723154676483434, + "grad_norm": 0.9131811261177063, + "learning_rate": 0.00016184667540192592, + "loss": 0.1902, + "step": 36690 + }, + { + "epoch": 0.5724714544206652, + "grad_norm": 0.21330870687961578, + "learning_rate": 0.00016183627628377114, + "loss": 0.2907, + "step": 36700 + }, + { + "epoch": 0.5726274411929868, + "grad_norm": 0.7145895957946777, + "learning_rate": 0.00016182587716561636, + "loss": 0.2574, + "step": 36710 + }, + { + "epoch": 0.5727834279653086, + "grad_norm": 1.0699433088302612, + "learning_rate": 0.00016181547804746158, + "loss": 0.1675, + "step": 36720 + }, + { + "epoch": 0.5729394147376302, + "grad_norm": 0.5066679120063782, + "learning_rate": 0.0001618050789293068, + "loss": 0.4487, + "step": 36730 + }, + { + "epoch": 0.573095401509952, + "grad_norm": 0.7482333183288574, + "learning_rate": 0.00016179467981115202, + "loss": 0.4101, + "step": 36740 + }, + { + "epoch": 0.5732513882822736, + "grad_norm": 0.15925532579421997, + "learning_rate": 0.00016178428069299724, + "loss": 0.2825, + "step": 36750 + }, + { + "epoch": 0.5734073750545954, + "grad_norm": 0.95689457654953, + "learning_rate": 0.00016177388157484246, + "loss": 0.2754, + "step": 36760 + }, + { + "epoch": 0.573563361826917, + "grad_norm": 2.15043568611145, + "learning_rate": 0.00016176348245668768, + "loss": 0.2073, + "step": 36770 + }, + { + "epoch": 0.5737193485992388, + "grad_norm": 2.651520252227783, + "learning_rate": 0.0001617530833385329, + "loss": 0.2517, + "step": 36780 + }, + { + "epoch": 0.5738753353715605, + "grad_norm": 2.1156702041625977, + "learning_rate": 0.00016174268422037812, + "loss": 0.3029, + "step": 36790 + }, + { + "epoch": 0.5740313221438822, + "grad_norm": 3.6200709342956543, + "learning_rate": 0.00016173228510222334, + "loss": 0.4174, + "step": 36800 + }, + { + "epoch": 0.574187308916204, + "grad_norm": 1.447937250137329, + "learning_rate": 0.00016172188598406856, + "loss": 0.1763, + "step": 36810 + }, + { + "epoch": 0.5743432956885256, + "grad_norm": 1.4020819664001465, + "learning_rate": 0.00016171148686591378, + "loss": 0.3994, + "step": 36820 + }, + { + "epoch": 0.5744992824608474, + "grad_norm": 1.4079807996749878, + "learning_rate": 0.000161701087747759, + "loss": 0.2051, + "step": 36830 + }, + { + "epoch": 0.574655269233169, + "grad_norm": 0.05268567427992821, + "learning_rate": 0.00016169068862960422, + "loss": 0.1773, + "step": 36840 + }, + { + "epoch": 0.5748112560054908, + "grad_norm": 2.248767614364624, + "learning_rate": 0.00016168028951144944, + "loss": 0.2534, + "step": 36850 + }, + { + "epoch": 0.5749672427778124, + "grad_norm": 0.3726302981376648, + "learning_rate": 0.00016166989039329466, + "loss": 0.328, + "step": 36860 + }, + { + "epoch": 0.5751232295501342, + "grad_norm": 1.0772110223770142, + "learning_rate": 0.00016165949127513988, + "loss": 0.2069, + "step": 36870 + }, + { + "epoch": 0.5752792163224558, + "grad_norm": 0.11339398473501205, + "learning_rate": 0.0001616490921569851, + "loss": 0.1827, + "step": 36880 + }, + { + "epoch": 0.5754352030947776, + "grad_norm": 1.2981103658676147, + "learning_rate": 0.00016163869303883032, + "loss": 0.1378, + "step": 36890 + }, + { + "epoch": 0.5755911898670992, + "grad_norm": 0.3302285075187683, + "learning_rate": 0.00016162829392067554, + "loss": 0.2311, + "step": 36900 + }, + { + "epoch": 0.575747176639421, + "grad_norm": 0.10168436169624329, + "learning_rate": 0.00016161789480252076, + "loss": 0.1678, + "step": 36910 + }, + { + "epoch": 0.5759031634117426, + "grad_norm": 1.6778357028961182, + "learning_rate": 0.00016160749568436598, + "loss": 0.2385, + "step": 36920 + }, + { + "epoch": 0.5760591501840644, + "grad_norm": 8.1710205078125, + "learning_rate": 0.0001615970965662112, + "loss": 0.3287, + "step": 36930 + }, + { + "epoch": 0.5762151369563862, + "grad_norm": 2.6481053829193115, + "learning_rate": 0.00016158669744805641, + "loss": 0.3119, + "step": 36940 + }, + { + "epoch": 0.5763711237287078, + "grad_norm": 0.3120105266571045, + "learning_rate": 0.00016157629832990163, + "loss": 0.1985, + "step": 36950 + }, + { + "epoch": 0.5765271105010296, + "grad_norm": 2.7221410274505615, + "learning_rate": 0.00016156589921174685, + "loss": 0.3697, + "step": 36960 + }, + { + "epoch": 0.5766830972733512, + "grad_norm": 0.34397706389427185, + "learning_rate": 0.00016155550009359207, + "loss": 0.2162, + "step": 36970 + }, + { + "epoch": 0.576839084045673, + "grad_norm": 4.6421003341674805, + "learning_rate": 0.0001615451009754373, + "loss": 0.2892, + "step": 36980 + }, + { + "epoch": 0.5769950708179946, + "grad_norm": 1.5395855903625488, + "learning_rate": 0.0001615347018572825, + "loss": 0.1258, + "step": 36990 + }, + { + "epoch": 0.5771510575903164, + "grad_norm": 0.4586782455444336, + "learning_rate": 0.00016152430273912773, + "loss": 0.1752, + "step": 37000 + }, + { + "epoch": 0.577307044362638, + "grad_norm": 0.4277658760547638, + "learning_rate": 0.00016151390362097295, + "loss": 0.1722, + "step": 37010 + }, + { + "epoch": 0.5774630311349598, + "grad_norm": 1.0282249450683594, + "learning_rate": 0.00016150350450281817, + "loss": 0.1361, + "step": 37020 + }, + { + "epoch": 0.5776190179072814, + "grad_norm": 0.2496921867132187, + "learning_rate": 0.0001614931053846634, + "loss": 0.0954, + "step": 37030 + }, + { + "epoch": 0.5777750046796032, + "grad_norm": 1.7900562286376953, + "learning_rate": 0.0001614827062665086, + "loss": 0.1284, + "step": 37040 + }, + { + "epoch": 0.5779309914519248, + "grad_norm": 0.6629091501235962, + "learning_rate": 0.00016147230714835383, + "loss": 0.2657, + "step": 37050 + }, + { + "epoch": 0.5780869782242466, + "grad_norm": 2.502882242202759, + "learning_rate": 0.00016146190803019905, + "loss": 0.207, + "step": 37060 + }, + { + "epoch": 0.5782429649965682, + "grad_norm": 1.486069917678833, + "learning_rate": 0.00016145150891204427, + "loss": 0.1232, + "step": 37070 + }, + { + "epoch": 0.57839895176889, + "grad_norm": 0.13543163239955902, + "learning_rate": 0.0001614411097938895, + "loss": 0.3245, + "step": 37080 + }, + { + "epoch": 0.5785549385412118, + "grad_norm": 3.9703898429870605, + "learning_rate": 0.0001614307106757347, + "loss": 0.3104, + "step": 37090 + }, + { + "epoch": 0.5787109253135334, + "grad_norm": 3.2236313819885254, + "learning_rate": 0.00016142031155757993, + "loss": 0.1739, + "step": 37100 + }, + { + "epoch": 0.5788669120858552, + "grad_norm": 0.3854866325855255, + "learning_rate": 0.00016140991243942515, + "loss": 0.084, + "step": 37110 + }, + { + "epoch": 0.5790228988581768, + "grad_norm": 0.1981169581413269, + "learning_rate": 0.00016139951332127037, + "loss": 0.2367, + "step": 37120 + }, + { + "epoch": 0.5791788856304986, + "grad_norm": 2.7315785884857178, + "learning_rate": 0.0001613891142031156, + "loss": 0.1816, + "step": 37130 + }, + { + "epoch": 0.5793348724028202, + "grad_norm": 0.48251378536224365, + "learning_rate": 0.0001613787150849608, + "loss": 0.2123, + "step": 37140 + }, + { + "epoch": 0.579490859175142, + "grad_norm": 5.0431413650512695, + "learning_rate": 0.00016136831596680603, + "loss": 0.2489, + "step": 37150 + }, + { + "epoch": 0.5796468459474636, + "grad_norm": 0.81981360912323, + "learning_rate": 0.00016135791684865125, + "loss": 0.5235, + "step": 37160 + }, + { + "epoch": 0.5798028327197854, + "grad_norm": 0.5701965689659119, + "learning_rate": 0.00016134751773049647, + "loss": 0.1709, + "step": 37170 + }, + { + "epoch": 0.579958819492107, + "grad_norm": 1.3905616998672485, + "learning_rate": 0.00016133711861234169, + "loss": 0.2641, + "step": 37180 + }, + { + "epoch": 0.5801148062644288, + "grad_norm": 2.281285047531128, + "learning_rate": 0.0001613267194941869, + "loss": 0.3915, + "step": 37190 + }, + { + "epoch": 0.5802707930367504, + "grad_norm": 1.2038309574127197, + "learning_rate": 0.00016131632037603213, + "loss": 0.0881, + "step": 37200 + }, + { + "epoch": 0.5804267798090722, + "grad_norm": 0.3096400201320648, + "learning_rate": 0.00016130592125787735, + "loss": 0.1341, + "step": 37210 + }, + { + "epoch": 0.5805827665813938, + "grad_norm": 0.10219905525445938, + "learning_rate": 0.00016129552213972256, + "loss": 0.0867, + "step": 37220 + }, + { + "epoch": 0.5807387533537156, + "grad_norm": 0.13869646191596985, + "learning_rate": 0.00016128512302156778, + "loss": 0.1621, + "step": 37230 + }, + { + "epoch": 0.5808947401260373, + "grad_norm": 0.7992938160896301, + "learning_rate": 0.000161274723903413, + "loss": 0.275, + "step": 37240 + }, + { + "epoch": 0.581050726898359, + "grad_norm": 0.2635972201824188, + "learning_rate": 0.00016126432478525822, + "loss": 0.2429, + "step": 37250 + }, + { + "epoch": 0.5812067136706808, + "grad_norm": 1.7211238145828247, + "learning_rate": 0.00016125392566710344, + "loss": 0.1541, + "step": 37260 + }, + { + "epoch": 0.5813627004430024, + "grad_norm": 2.2448506355285645, + "learning_rate": 0.00016124352654894866, + "loss": 0.3689, + "step": 37270 + }, + { + "epoch": 0.5815186872153242, + "grad_norm": 1.238708257675171, + "learning_rate": 0.00016123312743079388, + "loss": 0.3292, + "step": 37280 + }, + { + "epoch": 0.5816746739876458, + "grad_norm": 0.5670503377914429, + "learning_rate": 0.0001612227283126391, + "loss": 0.1151, + "step": 37290 + }, + { + "epoch": 0.5818306607599676, + "grad_norm": 2.287304401397705, + "learning_rate": 0.00016121232919448432, + "loss": 0.3558, + "step": 37300 + }, + { + "epoch": 0.5819866475322892, + "grad_norm": 0.8009629845619202, + "learning_rate": 0.00016120193007632954, + "loss": 0.1861, + "step": 37310 + }, + { + "epoch": 0.582142634304611, + "grad_norm": 1.1901466846466064, + "learning_rate": 0.00016119153095817476, + "loss": 0.089, + "step": 37320 + }, + { + "epoch": 0.5822986210769326, + "grad_norm": 1.7198141813278198, + "learning_rate": 0.00016118113184001998, + "loss": 0.2931, + "step": 37330 + }, + { + "epoch": 0.5824546078492544, + "grad_norm": 0.8832749724388123, + "learning_rate": 0.0001611707327218652, + "loss": 0.2342, + "step": 37340 + }, + { + "epoch": 0.582610594621576, + "grad_norm": 1.0623290538787842, + "learning_rate": 0.00016116033360371042, + "loss": 0.2317, + "step": 37350 + }, + { + "epoch": 0.5827665813938978, + "grad_norm": 0.5173699855804443, + "learning_rate": 0.00016114993448555564, + "loss": 0.1299, + "step": 37360 + }, + { + "epoch": 0.5829225681662195, + "grad_norm": 0.05118720978498459, + "learning_rate": 0.00016113953536740086, + "loss": 0.0786, + "step": 37370 + }, + { + "epoch": 0.5830785549385412, + "grad_norm": 2.0060107707977295, + "learning_rate": 0.00016112913624924608, + "loss": 0.266, + "step": 37380 + }, + { + "epoch": 0.5832345417108629, + "grad_norm": 5.121693134307861, + "learning_rate": 0.0001611187371310913, + "loss": 0.1824, + "step": 37390 + }, + { + "epoch": 0.5833905284831846, + "grad_norm": 0.3568089008331299, + "learning_rate": 0.00016110833801293652, + "loss": 0.3189, + "step": 37400 + }, + { + "epoch": 0.5835465152555064, + "grad_norm": 0.8631492853164673, + "learning_rate": 0.00016109793889478174, + "loss": 0.1724, + "step": 37410 + }, + { + "epoch": 0.583702502027828, + "grad_norm": 1.9214404821395874, + "learning_rate": 0.00016108753977662693, + "loss": 0.2094, + "step": 37420 + }, + { + "epoch": 0.5838584888001498, + "grad_norm": 1.9055646657943726, + "learning_rate": 0.00016107714065847218, + "loss": 0.1258, + "step": 37430 + }, + { + "epoch": 0.5840144755724714, + "grad_norm": 0.6614194512367249, + "learning_rate": 0.00016106674154031737, + "loss": 0.1966, + "step": 37440 + }, + { + "epoch": 0.5841704623447932, + "grad_norm": 0.286883682012558, + "learning_rate": 0.00016105634242216262, + "loss": 0.588, + "step": 37450 + }, + { + "epoch": 0.5843264491171148, + "grad_norm": 0.5599405765533447, + "learning_rate": 0.0001610459433040078, + "loss": 0.3414, + "step": 37460 + }, + { + "epoch": 0.5844824358894366, + "grad_norm": 1.5908915996551514, + "learning_rate": 0.00016103554418585306, + "loss": 0.1264, + "step": 37470 + }, + { + "epoch": 0.5846384226617583, + "grad_norm": 1.0729930400848389, + "learning_rate": 0.00016102514506769825, + "loss": 0.4815, + "step": 37480 + }, + { + "epoch": 0.58479440943408, + "grad_norm": 0.8912356495857239, + "learning_rate": 0.0001610147459495435, + "loss": 0.2092, + "step": 37490 + }, + { + "epoch": 0.5849503962064017, + "grad_norm": 0.24116197228431702, + "learning_rate": 0.0001610043468313887, + "loss": 0.1552, + "step": 37500 + }, + { + "epoch": 0.5851063829787234, + "grad_norm": 2.245192527770996, + "learning_rate": 0.00016099394771323393, + "loss": 0.0553, + "step": 37510 + }, + { + "epoch": 0.5852623697510451, + "grad_norm": 0.11538795381784439, + "learning_rate": 0.00016098354859507913, + "loss": 0.0889, + "step": 37520 + }, + { + "epoch": 0.5854183565233668, + "grad_norm": 1.0372689962387085, + "learning_rate": 0.00016097314947692437, + "loss": 0.3909, + "step": 37530 + }, + { + "epoch": 0.5855743432956885, + "grad_norm": 1.1946550607681274, + "learning_rate": 0.00016096275035876957, + "loss": 0.2167, + "step": 37540 + }, + { + "epoch": 0.5857303300680102, + "grad_norm": 1.3474591970443726, + "learning_rate": 0.0001609523512406148, + "loss": 0.1561, + "step": 37550 + }, + { + "epoch": 0.585886316840332, + "grad_norm": 0.008936616592109203, + "learning_rate": 0.00016094195212246, + "loss": 0.3991, + "step": 37560 + }, + { + "epoch": 0.5860423036126536, + "grad_norm": 1.8822320699691772, + "learning_rate": 0.00016093155300430525, + "loss": 0.2438, + "step": 37570 + }, + { + "epoch": 0.5861982903849754, + "grad_norm": 1.1707524061203003, + "learning_rate": 0.00016092115388615044, + "loss": 0.308, + "step": 37580 + }, + { + "epoch": 0.586354277157297, + "grad_norm": 1.679641842842102, + "learning_rate": 0.0001609107547679957, + "loss": 0.159, + "step": 37590 + }, + { + "epoch": 0.5865102639296188, + "grad_norm": 1.6885284185409546, + "learning_rate": 0.00016090035564984088, + "loss": 0.157, + "step": 37600 + }, + { + "epoch": 0.5866662507019405, + "grad_norm": 0.3876996636390686, + "learning_rate": 0.00016088995653168613, + "loss": 0.4067, + "step": 37610 + }, + { + "epoch": 0.5868222374742622, + "grad_norm": 1.387480616569519, + "learning_rate": 0.00016087955741353132, + "loss": 0.1898, + "step": 37620 + }, + { + "epoch": 0.5869782242465839, + "grad_norm": 0.03653861582279205, + "learning_rate": 0.00016086915829537657, + "loss": 0.1652, + "step": 37630 + }, + { + "epoch": 0.5871342110189056, + "grad_norm": 0.6258412003517151, + "learning_rate": 0.00016085875917722176, + "loss": 0.2558, + "step": 37640 + }, + { + "epoch": 0.5872901977912273, + "grad_norm": 0.10519164055585861, + "learning_rate": 0.000160848360059067, + "loss": 0.243, + "step": 37650 + }, + { + "epoch": 0.587446184563549, + "grad_norm": 0.7370131015777588, + "learning_rate": 0.0001608379609409122, + "loss": 0.2343, + "step": 37660 + }, + { + "epoch": 0.5876021713358707, + "grad_norm": 0.6884517073631287, + "learning_rate": 0.00016082756182275745, + "loss": 0.1847, + "step": 37670 + }, + { + "epoch": 0.5877581581081924, + "grad_norm": 0.6326120495796204, + "learning_rate": 0.00016081716270460264, + "loss": 0.2258, + "step": 37680 + }, + { + "epoch": 0.5879141448805141, + "grad_norm": 7.697889804840088, + "learning_rate": 0.0001608067635864479, + "loss": 0.2068, + "step": 37690 + }, + { + "epoch": 0.5880701316528358, + "grad_norm": 1.8753420114517212, + "learning_rate": 0.00016079636446829308, + "loss": 0.3087, + "step": 37700 + }, + { + "epoch": 0.5882261184251576, + "grad_norm": 0.8591625094413757, + "learning_rate": 0.00016078596535013833, + "loss": 0.1082, + "step": 37710 + }, + { + "epoch": 0.5883821051974792, + "grad_norm": 0.6124423146247864, + "learning_rate": 0.00016077556623198352, + "loss": 0.1536, + "step": 37720 + }, + { + "epoch": 0.588538091969801, + "grad_norm": 0.1400165557861328, + "learning_rate": 0.00016076516711382877, + "loss": 0.145, + "step": 37730 + }, + { + "epoch": 0.5886940787421227, + "grad_norm": 1.9879004955291748, + "learning_rate": 0.00016075476799567396, + "loss": 0.4986, + "step": 37740 + }, + { + "epoch": 0.5888500655144444, + "grad_norm": 0.08687327802181244, + "learning_rate": 0.0001607443688775192, + "loss": 0.3365, + "step": 37750 + }, + { + "epoch": 0.5890060522867661, + "grad_norm": 0.3442278206348419, + "learning_rate": 0.0001607339697593644, + "loss": 0.2274, + "step": 37760 + }, + { + "epoch": 0.5891620390590878, + "grad_norm": 0.06124915927648544, + "learning_rate": 0.00016072357064120965, + "loss": 0.1915, + "step": 37770 + }, + { + "epoch": 0.5893180258314095, + "grad_norm": 1.0996816158294678, + "learning_rate": 0.00016071317152305484, + "loss": 0.1573, + "step": 37780 + }, + { + "epoch": 0.5894740126037312, + "grad_norm": 0.14738436043262482, + "learning_rate": 0.00016070277240490008, + "loss": 0.2524, + "step": 37790 + }, + { + "epoch": 0.5896299993760529, + "grad_norm": 2.3905467987060547, + "learning_rate": 0.00016069237328674528, + "loss": 0.1838, + "step": 37800 + }, + { + "epoch": 0.5897859861483746, + "grad_norm": 2.0800178050994873, + "learning_rate": 0.00016068197416859052, + "loss": 0.3749, + "step": 37810 + }, + { + "epoch": 0.5899419729206963, + "grad_norm": 3.254204511642456, + "learning_rate": 0.00016067157505043574, + "loss": 0.3777, + "step": 37820 + }, + { + "epoch": 0.590097959693018, + "grad_norm": 2.8447892665863037, + "learning_rate": 0.00016066117593228096, + "loss": 0.5338, + "step": 37830 + }, + { + "epoch": 0.5902539464653397, + "grad_norm": 0.624733030796051, + "learning_rate": 0.00016065077681412618, + "loss": 0.2331, + "step": 37840 + }, + { + "epoch": 0.5904099332376614, + "grad_norm": 0.6863507628440857, + "learning_rate": 0.0001606403776959714, + "loss": 0.1747, + "step": 37850 + }, + { + "epoch": 0.5905659200099832, + "grad_norm": 0.6785943508148193, + "learning_rate": 0.00016062997857781662, + "loss": 0.2064, + "step": 37860 + }, + { + "epoch": 0.5907219067823049, + "grad_norm": 0.5739938616752625, + "learning_rate": 0.00016061957945966181, + "loss": 0.2368, + "step": 37870 + }, + { + "epoch": 0.5908778935546266, + "grad_norm": 0.0616387277841568, + "learning_rate": 0.00016060918034150706, + "loss": 0.2224, + "step": 37880 + }, + { + "epoch": 0.5910338803269483, + "grad_norm": 2.745957136154175, + "learning_rate": 0.00016059878122335225, + "loss": 0.3161, + "step": 37890 + }, + { + "epoch": 0.59118986709927, + "grad_norm": 2.52472186088562, + "learning_rate": 0.0001605883821051975, + "loss": 0.1713, + "step": 37900 + }, + { + "epoch": 0.5913458538715917, + "grad_norm": 1.5691877603530884, + "learning_rate": 0.0001605779829870427, + "loss": 0.1402, + "step": 37910 + }, + { + "epoch": 0.5915018406439134, + "grad_norm": 1.848261833190918, + "learning_rate": 0.00016056758386888794, + "loss": 0.2305, + "step": 37920 + }, + { + "epoch": 0.5916578274162351, + "grad_norm": 0.36622482538223267, + "learning_rate": 0.00016055718475073313, + "loss": 0.1834, + "step": 37930 + }, + { + "epoch": 0.5918138141885568, + "grad_norm": 0.21201446652412415, + "learning_rate": 0.00016054678563257838, + "loss": 0.4682, + "step": 37940 + }, + { + "epoch": 0.5919698009608785, + "grad_norm": 0.38921502232551575, + "learning_rate": 0.00016053638651442357, + "loss": 0.2892, + "step": 37950 + }, + { + "epoch": 0.5921257877332002, + "grad_norm": 0.20777840912342072, + "learning_rate": 0.00016052598739626882, + "loss": 0.104, + "step": 37960 + }, + { + "epoch": 0.5922817745055219, + "grad_norm": 0.7090150713920593, + "learning_rate": 0.000160515588278114, + "loss": 0.2592, + "step": 37970 + }, + { + "epoch": 0.5924377612778436, + "grad_norm": 1.9939709901809692, + "learning_rate": 0.00016050518915995926, + "loss": 0.247, + "step": 37980 + }, + { + "epoch": 0.5925937480501653, + "grad_norm": 0.8696603178977966, + "learning_rate": 0.00016049479004180445, + "loss": 0.301, + "step": 37990 + }, + { + "epoch": 0.592749734822487, + "grad_norm": 0.4119095206260681, + "learning_rate": 0.0001604843909236497, + "loss": 0.2011, + "step": 38000 + }, + { + "epoch": 0.5929057215948088, + "grad_norm": 0.060975607484579086, + "learning_rate": 0.0001604739918054949, + "loss": 0.2286, + "step": 38010 + }, + { + "epoch": 0.5930617083671305, + "grad_norm": 0.7039875388145447, + "learning_rate": 0.00016046359268734014, + "loss": 0.0999, + "step": 38020 + }, + { + "epoch": 0.5932176951394522, + "grad_norm": 0.28402045369148254, + "learning_rate": 0.00016045319356918533, + "loss": 0.1835, + "step": 38030 + }, + { + "epoch": 0.5933736819117739, + "grad_norm": 2.217988967895508, + "learning_rate": 0.00016044279445103058, + "loss": 0.0785, + "step": 38040 + }, + { + "epoch": 0.5935296686840956, + "grad_norm": 1.4856069087982178, + "learning_rate": 0.00016043239533287577, + "loss": 0.2528, + "step": 38050 + }, + { + "epoch": 0.5936856554564173, + "grad_norm": 0.9875519871711731, + "learning_rate": 0.00016042199621472101, + "loss": 0.4445, + "step": 38060 + }, + { + "epoch": 0.593841642228739, + "grad_norm": 2.773216724395752, + "learning_rate": 0.0001604115970965662, + "loss": 0.5643, + "step": 38070 + }, + { + "epoch": 0.5939976290010607, + "grad_norm": 1.7584549188613892, + "learning_rate": 0.00016040119797841145, + "loss": 0.1446, + "step": 38080 + }, + { + "epoch": 0.5941536157733824, + "grad_norm": 1.2456908226013184, + "learning_rate": 0.00016039079886025665, + "loss": 0.1896, + "step": 38090 + }, + { + "epoch": 0.5943096025457041, + "grad_norm": 0.021098516881465912, + "learning_rate": 0.0001603803997421019, + "loss": 0.0876, + "step": 38100 + }, + { + "epoch": 0.5944655893180258, + "grad_norm": 1.8131749629974365, + "learning_rate": 0.00016037000062394709, + "loss": 0.1639, + "step": 38110 + }, + { + "epoch": 0.5946215760903475, + "grad_norm": 0.41819462180137634, + "learning_rate": 0.00016035960150579233, + "loss": 0.1311, + "step": 38120 + }, + { + "epoch": 0.5947775628626693, + "grad_norm": 0.6691219806671143, + "learning_rate": 0.00016034920238763753, + "loss": 0.1625, + "step": 38130 + }, + { + "epoch": 0.5949335496349909, + "grad_norm": 0.07854852080345154, + "learning_rate": 0.00016033880326948277, + "loss": 0.3177, + "step": 38140 + }, + { + "epoch": 0.5950895364073127, + "grad_norm": 0.7819689512252808, + "learning_rate": 0.00016032840415132796, + "loss": 0.4225, + "step": 38150 + }, + { + "epoch": 0.5952455231796344, + "grad_norm": 5.3962836265563965, + "learning_rate": 0.0001603180050331732, + "loss": 0.123, + "step": 38160 + }, + { + "epoch": 0.5954015099519561, + "grad_norm": 0.7460303902626038, + "learning_rate": 0.0001603076059150184, + "loss": 0.2182, + "step": 38170 + }, + { + "epoch": 0.5955574967242778, + "grad_norm": 0.26433685421943665, + "learning_rate": 0.00016029720679686365, + "loss": 0.1894, + "step": 38180 + }, + { + "epoch": 0.5957134834965995, + "grad_norm": 0.8701320290565491, + "learning_rate": 0.00016028680767870884, + "loss": 0.2137, + "step": 38190 + }, + { + "epoch": 0.5958694702689212, + "grad_norm": 3.9774930477142334, + "learning_rate": 0.0001602764085605541, + "loss": 0.4637, + "step": 38200 + }, + { + "epoch": 0.5960254570412429, + "grad_norm": 0.5801145434379578, + "learning_rate": 0.00016026600944239928, + "loss": 0.3036, + "step": 38210 + }, + { + "epoch": 0.5961814438135646, + "grad_norm": 1.1890723705291748, + "learning_rate": 0.00016025561032424453, + "loss": 0.0988, + "step": 38220 + }, + { + "epoch": 0.5963374305858863, + "grad_norm": 0.18521073460578918, + "learning_rate": 0.00016024521120608972, + "loss": 0.1529, + "step": 38230 + }, + { + "epoch": 0.596493417358208, + "grad_norm": 2.2602827548980713, + "learning_rate": 0.00016023481208793497, + "loss": 0.2815, + "step": 38240 + }, + { + "epoch": 0.5966494041305297, + "grad_norm": 0.4060989022254944, + "learning_rate": 0.00016022441296978016, + "loss": 0.3466, + "step": 38250 + }, + { + "epoch": 0.5968053909028515, + "grad_norm": 1.4743244647979736, + "learning_rate": 0.0001602140138516254, + "loss": 0.3561, + "step": 38260 + }, + { + "epoch": 0.5969613776751731, + "grad_norm": 5.080365180969238, + "learning_rate": 0.0001602036147334706, + "loss": 0.2914, + "step": 38270 + }, + { + "epoch": 0.5971173644474949, + "grad_norm": 0.16149812936782837, + "learning_rate": 0.00016019321561531585, + "loss": 0.1871, + "step": 38280 + }, + { + "epoch": 0.5972733512198165, + "grad_norm": 1.7535364627838135, + "learning_rate": 0.00016018281649716104, + "loss": 0.4905, + "step": 38290 + }, + { + "epoch": 0.5974293379921383, + "grad_norm": 0.4103987216949463, + "learning_rate": 0.00016017241737900629, + "loss": 0.3813, + "step": 38300 + }, + { + "epoch": 0.59758532476446, + "grad_norm": 0.5315369367599487, + "learning_rate": 0.00016016201826085148, + "loss": 0.3897, + "step": 38310 + }, + { + "epoch": 0.5977413115367817, + "grad_norm": 3.4553463459014893, + "learning_rate": 0.0001601516191426967, + "loss": 0.2047, + "step": 38320 + }, + { + "epoch": 0.5978972983091034, + "grad_norm": 2.225468397140503, + "learning_rate": 0.00016014122002454192, + "loss": 0.4028, + "step": 38330 + }, + { + "epoch": 0.5980532850814251, + "grad_norm": 2.000885486602783, + "learning_rate": 0.00016013082090638714, + "loss": 0.391, + "step": 38340 + }, + { + "epoch": 0.5982092718537468, + "grad_norm": 0.7886672616004944, + "learning_rate": 0.00016012042178823236, + "loss": 0.2994, + "step": 38350 + }, + { + "epoch": 0.5983652586260685, + "grad_norm": 0.5279917120933533, + "learning_rate": 0.00016011002267007758, + "loss": 0.1855, + "step": 38360 + }, + { + "epoch": 0.5985212453983902, + "grad_norm": 0.9789942502975464, + "learning_rate": 0.0001600996235519228, + "loss": 0.2122, + "step": 38370 + }, + { + "epoch": 0.5986772321707119, + "grad_norm": 0.6289750933647156, + "learning_rate": 0.00016008922443376802, + "loss": 0.3737, + "step": 38380 + }, + { + "epoch": 0.5988332189430337, + "grad_norm": 1.3074733018875122, + "learning_rate": 0.00016007882531561324, + "loss": 0.2382, + "step": 38390 + }, + { + "epoch": 0.5989892057153553, + "grad_norm": 1.745284914970398, + "learning_rate": 0.00016006842619745846, + "loss": 0.3167, + "step": 38400 + }, + { + "epoch": 0.5991451924876771, + "grad_norm": 0.1694452464580536, + "learning_rate": 0.00016005802707930368, + "loss": 0.1632, + "step": 38410 + }, + { + "epoch": 0.5993011792599987, + "grad_norm": 1.55560302734375, + "learning_rate": 0.0001600476279611489, + "loss": 0.2371, + "step": 38420 + }, + { + "epoch": 0.5994571660323205, + "grad_norm": 1.044997215270996, + "learning_rate": 0.00016003722884299411, + "loss": 0.0896, + "step": 38430 + }, + { + "epoch": 0.5996131528046421, + "grad_norm": 0.25250110030174255, + "learning_rate": 0.00016002682972483933, + "loss": 0.3096, + "step": 38440 + }, + { + "epoch": 0.5997691395769639, + "grad_norm": 0.1631859689950943, + "learning_rate": 0.00016001643060668455, + "loss": 0.2004, + "step": 38450 + }, + { + "epoch": 0.5999251263492856, + "grad_norm": 0.34967318177223206, + "learning_rate": 0.00016000603148852977, + "loss": 0.1476, + "step": 38460 + }, + { + "epoch": 0.6000811131216073, + "grad_norm": 1.8604000806808472, + "learning_rate": 0.000159995632370375, + "loss": 0.1884, + "step": 38470 + }, + { + "epoch": 0.600237099893929, + "grad_norm": 1.626848578453064, + "learning_rate": 0.0001599852332522202, + "loss": 0.5452, + "step": 38480 + }, + { + "epoch": 0.6003930866662507, + "grad_norm": 0.41935989260673523, + "learning_rate": 0.00015997483413406543, + "loss": 0.3047, + "step": 38490 + }, + { + "epoch": 0.6005490734385724, + "grad_norm": 1.355086088180542, + "learning_rate": 0.00015996443501591065, + "loss": 0.1246, + "step": 38500 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.344487878656e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-38500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4fb3967b5ca3d4078c8998636b8dd0f814e6f76c --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc2d7dc864662d321cace18b8124afe9b9b6687faa53ae34b43fb66dad3767ef +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..4250fac550b7449d56716154db8b1064fb7acdd7 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60ffb832f450866682ffc3d570be9be96bceaa0473fcccee2221c087f422a6d7 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..f6a6c4fb76120a0b4edf7e99dfc4b58c465447be Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..8ae919c7bf3a3179232ec3d5e6fd482f66f14f2e Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..c75573ee66bd63613b4a9e189c46d08ddcdf3e26 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/trainer_state.json @@ -0,0 +1,27334 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.6083484120546577, + "eval_steps": 500, + "global_step": 39000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + }, + { + "epoch": 0.5071129968178698, + "grad_norm": 1.0075199604034424, + "learning_rate": 0.00016619350679062414, + "loss": 0.2569, + "step": 32510 + }, + { + "epoch": 0.5072689835901916, + "grad_norm": 1.3489066362380981, + "learning_rate": 0.0001661831076724694, + "loss": 0.2565, + "step": 32520 + }, + { + "epoch": 0.5074249703625132, + "grad_norm": 0.2534058094024658, + "learning_rate": 0.00016617270855431458, + "loss": 0.2292, + "step": 32530 + }, + { + "epoch": 0.507580957134835, + "grad_norm": 0.31723281741142273, + "learning_rate": 0.00016616230943615983, + "loss": 0.1467, + "step": 32540 + }, + { + "epoch": 0.5077369439071566, + "grad_norm": 0.29659849405288696, + "learning_rate": 0.00016615191031800502, + "loss": 0.2768, + "step": 32550 + }, + { + "epoch": 0.5078929306794784, + "grad_norm": 1.5246435403823853, + "learning_rate": 0.00016614151119985027, + "loss": 0.216, + "step": 32560 + }, + { + "epoch": 0.5080489174518001, + "grad_norm": 0.6026854515075684, + "learning_rate": 0.00016613111208169546, + "loss": 0.2277, + "step": 32570 + }, + { + "epoch": 0.5082049042241218, + "grad_norm": 0.41994890570640564, + "learning_rate": 0.0001661207129635407, + "loss": 0.2433, + "step": 32580 + }, + { + "epoch": 0.5083608909964435, + "grad_norm": 2.084599733352661, + "learning_rate": 0.00016611031384538593, + "loss": 0.4081, + "step": 32590 + }, + { + "epoch": 0.5085168777687652, + "grad_norm": 0.2945229709148407, + "learning_rate": 0.00016609991472723115, + "loss": 0.1329, + "step": 32600 + }, + { + "epoch": 0.508672864541087, + "grad_norm": 1.4981423616409302, + "learning_rate": 0.00016608951560907636, + "loss": 0.2855, + "step": 32610 + }, + { + "epoch": 0.5088288513134086, + "grad_norm": 0.961737871170044, + "learning_rate": 0.00016607911649092158, + "loss": 0.1654, + "step": 32620 + }, + { + "epoch": 0.5089848380857304, + "grad_norm": 1.49734628200531, + "learning_rate": 0.0001660687173727668, + "loss": 0.2081, + "step": 32630 + }, + { + "epoch": 0.509140824858052, + "grad_norm": 2.538135051727295, + "learning_rate": 0.00016605831825461202, + "loss": 0.3023, + "step": 32640 + }, + { + "epoch": 0.5092968116303738, + "grad_norm": 0.2981272041797638, + "learning_rate": 0.00016604791913645724, + "loss": 0.214, + "step": 32650 + }, + { + "epoch": 0.5094527984026954, + "grad_norm": 1.8863650560379028, + "learning_rate": 0.00016603752001830246, + "loss": 0.3363, + "step": 32660 + }, + { + "epoch": 0.5096087851750172, + "grad_norm": 0.022720983251929283, + "learning_rate": 0.00016602712090014768, + "loss": 0.2463, + "step": 32670 + }, + { + "epoch": 0.5097647719473388, + "grad_norm": 1.154977798461914, + "learning_rate": 0.0001660167217819929, + "loss": 0.4484, + "step": 32680 + }, + { + "epoch": 0.5099207587196606, + "grad_norm": 0.9470499157905579, + "learning_rate": 0.00016600632266383812, + "loss": 0.1196, + "step": 32690 + }, + { + "epoch": 0.5100767454919822, + "grad_norm": 0.19955401122570038, + "learning_rate": 0.00016599592354568334, + "loss": 0.1762, + "step": 32700 + }, + { + "epoch": 0.510232732264304, + "grad_norm": 0.4098282754421234, + "learning_rate": 0.00016598552442752856, + "loss": 0.2739, + "step": 32710 + }, + { + "epoch": 0.5103887190366257, + "grad_norm": 0.21697258949279785, + "learning_rate": 0.00016597512530937378, + "loss": 0.2555, + "step": 32720 + }, + { + "epoch": 0.5105447058089474, + "grad_norm": 4.894479751586914, + "learning_rate": 0.000165964726191219, + "loss": 0.3137, + "step": 32730 + }, + { + "epoch": 0.5107006925812692, + "grad_norm": 0.8396711349487305, + "learning_rate": 0.00016595432707306422, + "loss": 0.3105, + "step": 32740 + }, + { + "epoch": 0.5108566793535908, + "grad_norm": 1.5494674444198608, + "learning_rate": 0.00016594392795490944, + "loss": 0.2271, + "step": 32750 + }, + { + "epoch": 0.5110126661259126, + "grad_norm": 1.3409888744354248, + "learning_rate": 0.00016593352883675466, + "loss": 0.1796, + "step": 32760 + }, + { + "epoch": 0.5111686528982342, + "grad_norm": 1.7590911388397217, + "learning_rate": 0.00016592312971859988, + "loss": 0.2185, + "step": 32770 + }, + { + "epoch": 0.511324639670556, + "grad_norm": 0.7350002527236938, + "learning_rate": 0.0001659127306004451, + "loss": 0.1911, + "step": 32780 + }, + { + "epoch": 0.5114806264428776, + "grad_norm": 0.7176411747932434, + "learning_rate": 0.00016590233148229032, + "loss": 0.6918, + "step": 32790 + }, + { + "epoch": 0.5116366132151994, + "grad_norm": 2.5005321502685547, + "learning_rate": 0.00016589193236413554, + "loss": 0.2329, + "step": 32800 + }, + { + "epoch": 0.511792599987521, + "grad_norm": 1.3916250467300415, + "learning_rate": 0.00016588153324598076, + "loss": 0.2102, + "step": 32810 + }, + { + "epoch": 0.5119485867598428, + "grad_norm": 0.721671998500824, + "learning_rate": 0.00016587113412782598, + "loss": 0.2965, + "step": 32820 + }, + { + "epoch": 0.5121045735321644, + "grad_norm": 0.6120114326477051, + "learning_rate": 0.0001658607350096712, + "loss": 0.2661, + "step": 32830 + }, + { + "epoch": 0.5122605603044862, + "grad_norm": 0.29123347997665405, + "learning_rate": 0.00016585033589151642, + "loss": 0.2455, + "step": 32840 + }, + { + "epoch": 0.5124165470768078, + "grad_norm": 2.7300868034362793, + "learning_rate": 0.00016583993677336164, + "loss": 0.2049, + "step": 32850 + }, + { + "epoch": 0.5125725338491296, + "grad_norm": 1.801928997039795, + "learning_rate": 0.00016582953765520686, + "loss": 0.1725, + "step": 32860 + }, + { + "epoch": 0.5127285206214514, + "grad_norm": 0.6406471729278564, + "learning_rate": 0.00016581913853705208, + "loss": 0.164, + "step": 32870 + }, + { + "epoch": 0.512884507393773, + "grad_norm": 0.9627325534820557, + "learning_rate": 0.0001658087394188973, + "loss": 0.1728, + "step": 32880 + }, + { + "epoch": 0.5130404941660948, + "grad_norm": 1.5446962118148804, + "learning_rate": 0.00016579834030074251, + "loss": 0.2345, + "step": 32890 + }, + { + "epoch": 0.5131964809384164, + "grad_norm": 0.28092706203460693, + "learning_rate": 0.0001657879411825877, + "loss": 0.2124, + "step": 32900 + }, + { + "epoch": 0.5133524677107382, + "grad_norm": 1.7751421928405762, + "learning_rate": 0.00016577754206443295, + "loss": 0.2367, + "step": 32910 + }, + { + "epoch": 0.5135084544830598, + "grad_norm": 1.6964634656906128, + "learning_rate": 0.00016576714294627815, + "loss": 0.2361, + "step": 32920 + }, + { + "epoch": 0.5136644412553816, + "grad_norm": 0.06759823113679886, + "learning_rate": 0.0001657567438281234, + "loss": 0.2329, + "step": 32930 + }, + { + "epoch": 0.5138204280277032, + "grad_norm": 2.2519280910491943, + "learning_rate": 0.00016574634470996859, + "loss": 0.1079, + "step": 32940 + }, + { + "epoch": 0.513976414800025, + "grad_norm": 3.167576551437378, + "learning_rate": 0.00016573594559181383, + "loss": 0.1764, + "step": 32950 + }, + { + "epoch": 0.5141324015723466, + "grad_norm": 1.6734397411346436, + "learning_rate": 0.00016572554647365903, + "loss": 0.2878, + "step": 32960 + }, + { + "epoch": 0.5142883883446684, + "grad_norm": 0.35627374053001404, + "learning_rate": 0.00016571514735550427, + "loss": 0.2337, + "step": 32970 + }, + { + "epoch": 0.51444437511699, + "grad_norm": 1.6321039199829102, + "learning_rate": 0.00016570474823734946, + "loss": 0.2551, + "step": 32980 + }, + { + "epoch": 0.5146003618893118, + "grad_norm": 1.352982997894287, + "learning_rate": 0.0001656943491191947, + "loss": 0.4467, + "step": 32990 + }, + { + "epoch": 0.5147563486616334, + "grad_norm": 0.3558129072189331, + "learning_rate": 0.0001656839500010399, + "loss": 0.1913, + "step": 33000 + }, + { + "epoch": 0.5149123354339552, + "grad_norm": 0.041972286999225616, + "learning_rate": 0.00016567355088288515, + "loss": 0.1704, + "step": 33010 + }, + { + "epoch": 0.515068322206277, + "grad_norm": 1.4168939590454102, + "learning_rate": 0.00016566315176473034, + "loss": 0.4866, + "step": 33020 + }, + { + "epoch": 0.5152243089785986, + "grad_norm": 3.0042929649353027, + "learning_rate": 0.0001656527526465756, + "loss": 0.2473, + "step": 33030 + }, + { + "epoch": 0.5153802957509204, + "grad_norm": 1.027274489402771, + "learning_rate": 0.00016564235352842078, + "loss": 0.436, + "step": 33040 + }, + { + "epoch": 0.515536282523242, + "grad_norm": 1.2330195903778076, + "learning_rate": 0.00016563195441026603, + "loss": 0.4615, + "step": 33050 + }, + { + "epoch": 0.5156922692955638, + "grad_norm": 2.7145066261291504, + "learning_rate": 0.00016562155529211122, + "loss": 0.278, + "step": 33060 + }, + { + "epoch": 0.5158482560678854, + "grad_norm": 1.1273893117904663, + "learning_rate": 0.00016561115617395647, + "loss": 0.3767, + "step": 33070 + }, + { + "epoch": 0.5160042428402072, + "grad_norm": 2.466265916824341, + "learning_rate": 0.00016560075705580166, + "loss": 0.1879, + "step": 33080 + }, + { + "epoch": 0.5161602296125288, + "grad_norm": 0.3759515583515167, + "learning_rate": 0.0001655903579376469, + "loss": 0.4004, + "step": 33090 + }, + { + "epoch": 0.5163162163848506, + "grad_norm": 1.8821336030960083, + "learning_rate": 0.0001655799588194921, + "loss": 0.1622, + "step": 33100 + }, + { + "epoch": 0.5164722031571722, + "grad_norm": 0.8162096738815308, + "learning_rate": 0.00016556955970133735, + "loss": 0.1341, + "step": 33110 + }, + { + "epoch": 0.516628189929494, + "grad_norm": 0.008498801849782467, + "learning_rate": 0.00016555916058318254, + "loss": 0.2546, + "step": 33120 + }, + { + "epoch": 0.5167841767018156, + "grad_norm": 0.17536675930023193, + "learning_rate": 0.00016554876146502779, + "loss": 0.2934, + "step": 33130 + }, + { + "epoch": 0.5169401634741374, + "grad_norm": 3.539898633956909, + "learning_rate": 0.00016553836234687298, + "loss": 0.4959, + "step": 33140 + }, + { + "epoch": 0.517096150246459, + "grad_norm": 2.190887212753296, + "learning_rate": 0.00016552796322871823, + "loss": 0.2914, + "step": 33150 + }, + { + "epoch": 0.5172521370187808, + "grad_norm": 1.6388943195343018, + "learning_rate": 0.00016551756411056342, + "loss": 0.1317, + "step": 33160 + }, + { + "epoch": 0.5174081237911026, + "grad_norm": 0.33063387870788574, + "learning_rate": 0.00016550716499240866, + "loss": 0.4258, + "step": 33170 + }, + { + "epoch": 0.5175641105634242, + "grad_norm": 0.5168288946151733, + "learning_rate": 0.00016549676587425386, + "loss": 0.2548, + "step": 33180 + }, + { + "epoch": 0.517720097335746, + "grad_norm": 3.9570281505584717, + "learning_rate": 0.0001654863667560991, + "loss": 0.3179, + "step": 33190 + }, + { + "epoch": 0.5178760841080676, + "grad_norm": 1.127255916595459, + "learning_rate": 0.0001654759676379443, + "loss": 0.233, + "step": 33200 + }, + { + "epoch": 0.5180320708803894, + "grad_norm": 0.36299410462379456, + "learning_rate": 0.00016546556851978954, + "loss": 0.4614, + "step": 33210 + }, + { + "epoch": 0.518188057652711, + "grad_norm": 2.914668321609497, + "learning_rate": 0.00016545516940163474, + "loss": 0.1603, + "step": 33220 + }, + { + "epoch": 0.5183440444250328, + "grad_norm": 6.991094589233398, + "learning_rate": 0.00016544477028347998, + "loss": 0.2259, + "step": 33230 + }, + { + "epoch": 0.5185000311973544, + "grad_norm": 2.0594241619110107, + "learning_rate": 0.00016543437116532517, + "loss": 0.1885, + "step": 33240 + }, + { + "epoch": 0.5186560179696762, + "grad_norm": 0.5594693422317505, + "learning_rate": 0.00016542397204717042, + "loss": 0.4333, + "step": 33250 + }, + { + "epoch": 0.5188120047419978, + "grad_norm": 4.314854621887207, + "learning_rate": 0.00016541357292901561, + "loss": 0.2103, + "step": 33260 + }, + { + "epoch": 0.5189679915143196, + "grad_norm": 0.5609000325202942, + "learning_rate": 0.00016540317381086086, + "loss": 0.2498, + "step": 33270 + }, + { + "epoch": 0.5191239782866413, + "grad_norm": 3.680088758468628, + "learning_rate": 0.00016539277469270605, + "loss": 0.3465, + "step": 33280 + }, + { + "epoch": 0.519279965058963, + "grad_norm": 1.4993672370910645, + "learning_rate": 0.0001653823755745513, + "loss": 0.3272, + "step": 33290 + }, + { + "epoch": 0.5194359518312847, + "grad_norm": 0.18005084991455078, + "learning_rate": 0.0001653719764563965, + "loss": 0.1706, + "step": 33300 + }, + { + "epoch": 0.5195919386036064, + "grad_norm": 0.5105524659156799, + "learning_rate": 0.00016536157733824174, + "loss": 0.0937, + "step": 33310 + }, + { + "epoch": 0.5197479253759281, + "grad_norm": 1.2806589603424072, + "learning_rate": 0.00016535117822008693, + "loss": 0.1578, + "step": 33320 + }, + { + "epoch": 0.5199039121482498, + "grad_norm": 0.22015990316867828, + "learning_rate": 0.00016534077910193218, + "loss": 0.2081, + "step": 33330 + }, + { + "epoch": 0.5200598989205716, + "grad_norm": 0.34638991951942444, + "learning_rate": 0.00016533037998377737, + "loss": 0.2389, + "step": 33340 + }, + { + "epoch": 0.5202158856928932, + "grad_norm": 0.8709302544593811, + "learning_rate": 0.0001653199808656226, + "loss": 0.2112, + "step": 33350 + }, + { + "epoch": 0.520371872465215, + "grad_norm": 1.2309521436691284, + "learning_rate": 0.0001653095817474678, + "loss": 0.3448, + "step": 33360 + }, + { + "epoch": 0.5205278592375366, + "grad_norm": 1.3155115842819214, + "learning_rate": 0.00016529918262931303, + "loss": 0.1572, + "step": 33370 + }, + { + "epoch": 0.5206838460098584, + "grad_norm": 0.358751118183136, + "learning_rate": 0.00016528878351115825, + "loss": 0.3011, + "step": 33380 + }, + { + "epoch": 0.52083983278218, + "grad_norm": 3.800628662109375, + "learning_rate": 0.00016527838439300347, + "loss": 0.458, + "step": 33390 + }, + { + "epoch": 0.5209958195545018, + "grad_norm": 1.1253124475479126, + "learning_rate": 0.0001652679852748487, + "loss": 0.1238, + "step": 33400 + }, + { + "epoch": 0.5211518063268235, + "grad_norm": 0.46979108452796936, + "learning_rate": 0.0001652575861566939, + "loss": 0.0745, + "step": 33410 + }, + { + "epoch": 0.5213077930991452, + "grad_norm": 7.921492576599121, + "learning_rate": 0.00016524718703853913, + "loss": 0.4004, + "step": 33420 + }, + { + "epoch": 0.5214637798714669, + "grad_norm": 1.643127679824829, + "learning_rate": 0.00016523678792038435, + "loss": 0.2858, + "step": 33430 + }, + { + "epoch": 0.5216197666437886, + "grad_norm": 7.096631050109863, + "learning_rate": 0.0001652263888022296, + "loss": 0.4735, + "step": 33440 + }, + { + "epoch": 0.5217757534161103, + "grad_norm": 0.7280219197273254, + "learning_rate": 0.0001652159896840748, + "loss": 0.1841, + "step": 33450 + }, + { + "epoch": 0.521931740188432, + "grad_norm": 0.9415522813796997, + "learning_rate": 0.00016520559056592003, + "loss": 0.2169, + "step": 33460 + }, + { + "epoch": 0.5220877269607537, + "grad_norm": 2.264862060546875, + "learning_rate": 0.00016519519144776523, + "loss": 0.1902, + "step": 33470 + }, + { + "epoch": 0.5222437137330754, + "grad_norm": 2.4566822052001953, + "learning_rate": 0.00016518479232961047, + "loss": 0.3329, + "step": 33480 + }, + { + "epoch": 0.5223997005053972, + "grad_norm": 0.7330378890037537, + "learning_rate": 0.00016517439321145567, + "loss": 0.2341, + "step": 33490 + }, + { + "epoch": 0.5225556872777188, + "grad_norm": 1.6409542560577393, + "learning_rate": 0.0001651639940933009, + "loss": 0.3077, + "step": 33500 + }, + { + "epoch": 0.5227116740500406, + "grad_norm": 1.0647432804107666, + "learning_rate": 0.0001651535949751461, + "loss": 0.1935, + "step": 33510 + }, + { + "epoch": 0.5228676608223622, + "grad_norm": 0.507420539855957, + "learning_rate": 0.00016514319585699135, + "loss": 0.1314, + "step": 33520 + }, + { + "epoch": 0.523023647594684, + "grad_norm": 1.0064164400100708, + "learning_rate": 0.00016513279673883654, + "loss": 0.344, + "step": 33530 + }, + { + "epoch": 0.5231796343670057, + "grad_norm": 1.1936209201812744, + "learning_rate": 0.0001651223976206818, + "loss": 0.1948, + "step": 33540 + }, + { + "epoch": 0.5233356211393274, + "grad_norm": 0.04992926865816116, + "learning_rate": 0.00016511199850252698, + "loss": 0.1363, + "step": 33550 + }, + { + "epoch": 0.5234916079116491, + "grad_norm": 1.8256275653839111, + "learning_rate": 0.00016510159938437223, + "loss": 0.443, + "step": 33560 + }, + { + "epoch": 0.5236475946839708, + "grad_norm": 4.914908409118652, + "learning_rate": 0.00016509120026621742, + "loss": 0.4654, + "step": 33570 + }, + { + "epoch": 0.5238035814562925, + "grad_norm": 1.1625486612319946, + "learning_rate": 0.00016508080114806267, + "loss": 0.1974, + "step": 33580 + }, + { + "epoch": 0.5239595682286142, + "grad_norm": 6.6192522048950195, + "learning_rate": 0.00016507040202990786, + "loss": 0.2131, + "step": 33590 + }, + { + "epoch": 0.5241155550009359, + "grad_norm": 0.7369065880775452, + "learning_rate": 0.0001650600029117531, + "loss": 0.2371, + "step": 33600 + }, + { + "epoch": 0.5242715417732576, + "grad_norm": 1.5238152742385864, + "learning_rate": 0.0001650496037935983, + "loss": 0.1557, + "step": 33610 + }, + { + "epoch": 0.5244275285455793, + "grad_norm": 1.0418007373809814, + "learning_rate": 0.00016503920467544355, + "loss": 0.1878, + "step": 33620 + }, + { + "epoch": 0.524583515317901, + "grad_norm": 0.790117084980011, + "learning_rate": 0.00016502880555728874, + "loss": 0.2195, + "step": 33630 + }, + { + "epoch": 0.5247395020902228, + "grad_norm": 1.6712257862091064, + "learning_rate": 0.000165018406439134, + "loss": 0.1602, + "step": 33640 + }, + { + "epoch": 0.5248954888625444, + "grad_norm": 0.19236230850219727, + "learning_rate": 0.00016500800732097918, + "loss": 0.2526, + "step": 33650 + }, + { + "epoch": 0.5250514756348662, + "grad_norm": 1.3519701957702637, + "learning_rate": 0.00016499760820282443, + "loss": 0.4686, + "step": 33660 + }, + { + "epoch": 0.5252074624071879, + "grad_norm": 1.694342851638794, + "learning_rate": 0.00016498720908466962, + "loss": 0.1859, + "step": 33670 + }, + { + "epoch": 0.5253634491795096, + "grad_norm": 5.225239276885986, + "learning_rate": 0.00016497680996651487, + "loss": 0.2944, + "step": 33680 + }, + { + "epoch": 0.5255194359518313, + "grad_norm": 2.0208842754364014, + "learning_rate": 0.00016496641084836006, + "loss": 0.2421, + "step": 33690 + }, + { + "epoch": 0.525675422724153, + "grad_norm": 0.7954996228218079, + "learning_rate": 0.0001649560117302053, + "loss": 0.2853, + "step": 33700 + }, + { + "epoch": 0.5258314094964747, + "grad_norm": 2.296086072921753, + "learning_rate": 0.0001649456126120505, + "loss": 0.1333, + "step": 33710 + }, + { + "epoch": 0.5259873962687964, + "grad_norm": 1.1779128313064575, + "learning_rate": 0.00016493521349389574, + "loss": 0.1066, + "step": 33720 + }, + { + "epoch": 0.5261433830411181, + "grad_norm": 0.1756065934896469, + "learning_rate": 0.00016492481437574094, + "loss": 0.1352, + "step": 33730 + }, + { + "epoch": 0.5262993698134398, + "grad_norm": 0.13100725412368774, + "learning_rate": 0.00016491441525758618, + "loss": 0.2399, + "step": 33740 + }, + { + "epoch": 0.5264553565857615, + "grad_norm": 5.532008171081543, + "learning_rate": 0.00016490401613943138, + "loss": 0.2896, + "step": 33750 + }, + { + "epoch": 0.5266113433580832, + "grad_norm": 1.319886565208435, + "learning_rate": 0.00016489361702127662, + "loss": 0.3275, + "step": 33760 + }, + { + "epoch": 0.5267673301304049, + "grad_norm": 1.5550974607467651, + "learning_rate": 0.00016488321790312182, + "loss": 0.2677, + "step": 33770 + }, + { + "epoch": 0.5269233169027266, + "grad_norm": 1.8936737775802612, + "learning_rate": 0.00016487281878496706, + "loss": 0.1955, + "step": 33780 + }, + { + "epoch": 0.5270793036750484, + "grad_norm": 0.3653401732444763, + "learning_rate": 0.00016486241966681226, + "loss": 0.0723, + "step": 33790 + }, + { + "epoch": 0.52723529044737, + "grad_norm": 2.861341714859009, + "learning_rate": 0.00016485202054865747, + "loss": 0.2412, + "step": 33800 + }, + { + "epoch": 0.5273912772196918, + "grad_norm": 1.5291428565979004, + "learning_rate": 0.0001648416214305027, + "loss": 0.0871, + "step": 33810 + }, + { + "epoch": 0.5275472639920135, + "grad_norm": 1.0372581481933594, + "learning_rate": 0.00016483122231234791, + "loss": 0.4705, + "step": 33820 + }, + { + "epoch": 0.5277032507643352, + "grad_norm": 1.1943141222000122, + "learning_rate": 0.00016482082319419313, + "loss": 0.2848, + "step": 33830 + }, + { + "epoch": 0.5278592375366569, + "grad_norm": 1.9008225202560425, + "learning_rate": 0.00016481042407603835, + "loss": 0.139, + "step": 33840 + }, + { + "epoch": 0.5280152243089786, + "grad_norm": 2.132089138031006, + "learning_rate": 0.00016480002495788357, + "loss": 0.2119, + "step": 33850 + }, + { + "epoch": 0.5281712110813003, + "grad_norm": 0.24524426460266113, + "learning_rate": 0.0001647896258397288, + "loss": 0.1643, + "step": 33860 + }, + { + "epoch": 0.528327197853622, + "grad_norm": 1.6469637155532837, + "learning_rate": 0.000164779226721574, + "loss": 0.363, + "step": 33870 + }, + { + "epoch": 0.5284831846259437, + "grad_norm": 0.8767328858375549, + "learning_rate": 0.00016476882760341923, + "loss": 0.2632, + "step": 33880 + }, + { + "epoch": 0.5286391713982654, + "grad_norm": 0.06347586214542389, + "learning_rate": 0.00016475842848526445, + "loss": 0.3204, + "step": 33890 + }, + { + "epoch": 0.5287951581705871, + "grad_norm": 0.09782540798187256, + "learning_rate": 0.00016474802936710967, + "loss": 0.2211, + "step": 33900 + }, + { + "epoch": 0.5289511449429088, + "grad_norm": 3.2998859882354736, + "learning_rate": 0.0001647376302489549, + "loss": 0.265, + "step": 33910 + }, + { + "epoch": 0.5291071317152305, + "grad_norm": 0.43594226241111755, + "learning_rate": 0.0001647272311308001, + "loss": 0.194, + "step": 33920 + }, + { + "epoch": 0.5292631184875523, + "grad_norm": 1.5166605710983276, + "learning_rate": 0.00016471683201264533, + "loss": 0.2675, + "step": 33930 + }, + { + "epoch": 0.529419105259874, + "grad_norm": 0.6056640148162842, + "learning_rate": 0.00016470643289449055, + "loss": 0.0576, + "step": 33940 + }, + { + "epoch": 0.5295750920321957, + "grad_norm": 0.25410348176956177, + "learning_rate": 0.00016469603377633577, + "loss": 0.1124, + "step": 33950 + }, + { + "epoch": 0.5297310788045174, + "grad_norm": 1.770642876625061, + "learning_rate": 0.000164685634658181, + "loss": 0.3295, + "step": 33960 + }, + { + "epoch": 0.5298870655768391, + "grad_norm": 0.0607205331325531, + "learning_rate": 0.0001646752355400262, + "loss": 0.2369, + "step": 33970 + }, + { + "epoch": 0.5300430523491608, + "grad_norm": 0.5557095408439636, + "learning_rate": 0.00016466483642187143, + "loss": 0.1681, + "step": 33980 + }, + { + "epoch": 0.5301990391214825, + "grad_norm": 0.5192957520484924, + "learning_rate": 0.00016465443730371665, + "loss": 0.2649, + "step": 33990 + }, + { + "epoch": 0.5303550258938042, + "grad_norm": 0.04804835096001625, + "learning_rate": 0.00016464403818556187, + "loss": 0.2639, + "step": 34000 + }, + { + "epoch": 0.5305110126661259, + "grad_norm": 0.02673129364848137, + "learning_rate": 0.0001646336390674071, + "loss": 0.169, + "step": 34010 + }, + { + "epoch": 0.5306669994384476, + "grad_norm": 1.0084244012832642, + "learning_rate": 0.0001646232399492523, + "loss": 0.353, + "step": 34020 + }, + { + "epoch": 0.5308229862107693, + "grad_norm": 2.2202091217041016, + "learning_rate": 0.00016461284083109753, + "loss": 0.3206, + "step": 34030 + }, + { + "epoch": 0.530978972983091, + "grad_norm": 0.5573744773864746, + "learning_rate": 0.00016460244171294275, + "loss": 0.231, + "step": 34040 + }, + { + "epoch": 0.5311349597554127, + "grad_norm": 0.39700084924697876, + "learning_rate": 0.00016459204259478797, + "loss": 0.1412, + "step": 34050 + }, + { + "epoch": 0.5312909465277345, + "grad_norm": 2.582963228225708, + "learning_rate": 0.00016458164347663319, + "loss": 0.3334, + "step": 34060 + }, + { + "epoch": 0.5314469333000561, + "grad_norm": 0.22781169414520264, + "learning_rate": 0.0001645712443584784, + "loss": 0.2767, + "step": 34070 + }, + { + "epoch": 0.5316029200723779, + "grad_norm": 1.2051042318344116, + "learning_rate": 0.00016456084524032362, + "loss": 0.2045, + "step": 34080 + }, + { + "epoch": 0.5317589068446996, + "grad_norm": 0.42760100960731506, + "learning_rate": 0.00016455044612216884, + "loss": 0.1151, + "step": 34090 + }, + { + "epoch": 0.5319148936170213, + "grad_norm": 0.14440476894378662, + "learning_rate": 0.00016454004700401406, + "loss": 0.2386, + "step": 34100 + }, + { + "epoch": 0.532070880389343, + "grad_norm": 2.2777981758117676, + "learning_rate": 0.00016452964788585928, + "loss": 0.2009, + "step": 34110 + }, + { + "epoch": 0.5322268671616647, + "grad_norm": 0.9206979274749756, + "learning_rate": 0.0001645192487677045, + "loss": 0.2745, + "step": 34120 + }, + { + "epoch": 0.5323828539339864, + "grad_norm": 1.6947574615478516, + "learning_rate": 0.00016450884964954972, + "loss": 0.2584, + "step": 34130 + }, + { + "epoch": 0.5325388407063081, + "grad_norm": 0.401444673538208, + "learning_rate": 0.00016449845053139494, + "loss": 0.2218, + "step": 34140 + }, + { + "epoch": 0.5326948274786298, + "grad_norm": 0.08261553198099136, + "learning_rate": 0.00016448805141324016, + "loss": 0.2775, + "step": 34150 + }, + { + "epoch": 0.5328508142509515, + "grad_norm": 0.1017974391579628, + "learning_rate": 0.00016447765229508538, + "loss": 0.2095, + "step": 34160 + }, + { + "epoch": 0.5330068010232732, + "grad_norm": 1.3759571313858032, + "learning_rate": 0.0001644672531769306, + "loss": 0.2643, + "step": 34170 + }, + { + "epoch": 0.5331627877955949, + "grad_norm": 1.2654389142990112, + "learning_rate": 0.00016445685405877582, + "loss": 0.2949, + "step": 34180 + }, + { + "epoch": 0.5333187745679167, + "grad_norm": 1.5481843948364258, + "learning_rate": 0.00016444645494062104, + "loss": 0.1264, + "step": 34190 + }, + { + "epoch": 0.5334747613402383, + "grad_norm": 1.8094528913497925, + "learning_rate": 0.00016443605582246626, + "loss": 0.2727, + "step": 34200 + }, + { + "epoch": 0.5336307481125601, + "grad_norm": 2.224538564682007, + "learning_rate": 0.00016442565670431148, + "loss": 0.3096, + "step": 34210 + }, + { + "epoch": 0.5337867348848817, + "grad_norm": 0.6375226974487305, + "learning_rate": 0.0001644152575861567, + "loss": 0.2251, + "step": 34220 + }, + { + "epoch": 0.5339427216572035, + "grad_norm": 3.727106809616089, + "learning_rate": 0.00016440485846800192, + "loss": 0.4374, + "step": 34230 + }, + { + "epoch": 0.5340987084295252, + "grad_norm": 0.13345426321029663, + "learning_rate": 0.00016439445934984714, + "loss": 0.2011, + "step": 34240 + }, + { + "epoch": 0.5342546952018469, + "grad_norm": 2.1658668518066406, + "learning_rate": 0.00016438406023169236, + "loss": 0.2457, + "step": 34250 + }, + { + "epoch": 0.5344106819741686, + "grad_norm": 19.238407135009766, + "learning_rate": 0.00016437366111353758, + "loss": 0.2756, + "step": 34260 + }, + { + "epoch": 0.5345666687464903, + "grad_norm": 1.0292778015136719, + "learning_rate": 0.0001643632619953828, + "loss": 0.1646, + "step": 34270 + }, + { + "epoch": 0.534722655518812, + "grad_norm": 0.9372987747192383, + "learning_rate": 0.00016435286287722802, + "loss": 0.2762, + "step": 34280 + }, + { + "epoch": 0.5348786422911337, + "grad_norm": 0.3918002247810364, + "learning_rate": 0.00016434246375907324, + "loss": 0.1406, + "step": 34290 + }, + { + "epoch": 0.5350346290634554, + "grad_norm": 1.3518732786178589, + "learning_rate": 0.00016433206464091846, + "loss": 0.2992, + "step": 34300 + }, + { + "epoch": 0.5351906158357771, + "grad_norm": 0.73117595911026, + "learning_rate": 0.00016432166552276368, + "loss": 0.1172, + "step": 34310 + }, + { + "epoch": 0.5353466026080989, + "grad_norm": 0.06655958294868469, + "learning_rate": 0.0001643112664046089, + "loss": 0.1332, + "step": 34320 + }, + { + "epoch": 0.5355025893804205, + "grad_norm": 3.2587168216705322, + "learning_rate": 0.00016430086728645412, + "loss": 0.2052, + "step": 34330 + }, + { + "epoch": 0.5356585761527423, + "grad_norm": 2.6668431758880615, + "learning_rate": 0.00016429046816829934, + "loss": 0.1295, + "step": 34340 + }, + { + "epoch": 0.5358145629250639, + "grad_norm": 1.316080927848816, + "learning_rate": 0.00016428006905014456, + "loss": 0.2072, + "step": 34350 + }, + { + "epoch": 0.5359705496973857, + "grad_norm": 5.144528388977051, + "learning_rate": 0.00016426966993198977, + "loss": 0.1934, + "step": 34360 + }, + { + "epoch": 0.5361265364697073, + "grad_norm": 1.3529599905014038, + "learning_rate": 0.000164259270813835, + "loss": 0.2077, + "step": 34370 + }, + { + "epoch": 0.5362825232420291, + "grad_norm": 0.9132925868034363, + "learning_rate": 0.00016424887169568021, + "loss": 0.2404, + "step": 34380 + }, + { + "epoch": 0.5364385100143508, + "grad_norm": 1.520033597946167, + "learning_rate": 0.00016423847257752543, + "loss": 0.1765, + "step": 34390 + }, + { + "epoch": 0.5365944967866725, + "grad_norm": 2.343975305557251, + "learning_rate": 0.00016422807345937065, + "loss": 0.1959, + "step": 34400 + }, + { + "epoch": 0.5367504835589942, + "grad_norm": 0.5855118632316589, + "learning_rate": 0.00016421767434121587, + "loss": 0.0848, + "step": 34410 + }, + { + "epoch": 0.5369064703313159, + "grad_norm": 1.9108648300170898, + "learning_rate": 0.0001642072752230611, + "loss": 0.2406, + "step": 34420 + }, + { + "epoch": 0.5370624571036376, + "grad_norm": 1.0323792695999146, + "learning_rate": 0.0001641968761049063, + "loss": 0.2298, + "step": 34430 + }, + { + "epoch": 0.5372184438759593, + "grad_norm": 0.17435620725154877, + "learning_rate": 0.00016418647698675153, + "loss": 0.1396, + "step": 34440 + }, + { + "epoch": 0.537374430648281, + "grad_norm": 0.3526577353477478, + "learning_rate": 0.00016417607786859675, + "loss": 0.172, + "step": 34450 + }, + { + "epoch": 0.5375304174206027, + "grad_norm": 2.73685884475708, + "learning_rate": 0.00016416567875044197, + "loss": 0.1167, + "step": 34460 + }, + { + "epoch": 0.5376864041929245, + "grad_norm": 2.765693426132202, + "learning_rate": 0.0001641552796322872, + "loss": 0.1356, + "step": 34470 + }, + { + "epoch": 0.5378423909652461, + "grad_norm": 0.47628021240234375, + "learning_rate": 0.0001641448805141324, + "loss": 0.3104, + "step": 34480 + }, + { + "epoch": 0.5379983777375679, + "grad_norm": 1.1307590007781982, + "learning_rate": 0.00016413448139597763, + "loss": 0.3247, + "step": 34490 + }, + { + "epoch": 0.5381543645098895, + "grad_norm": 0.07136381417512894, + "learning_rate": 0.00016412408227782285, + "loss": 0.319, + "step": 34500 + }, + { + "epoch": 0.5383103512822113, + "grad_norm": 1.7850221395492554, + "learning_rate": 0.00016411368315966807, + "loss": 0.2515, + "step": 34510 + }, + { + "epoch": 0.5384663380545329, + "grad_norm": 1.3870742321014404, + "learning_rate": 0.0001641032840415133, + "loss": 0.256, + "step": 34520 + }, + { + "epoch": 0.5386223248268547, + "grad_norm": 1.1329221725463867, + "learning_rate": 0.0001640928849233585, + "loss": 0.2255, + "step": 34530 + }, + { + "epoch": 0.5387783115991764, + "grad_norm": 0.23236137628555298, + "learning_rate": 0.00016408248580520373, + "loss": 0.4053, + "step": 34540 + }, + { + "epoch": 0.5389342983714981, + "grad_norm": 3.350924253463745, + "learning_rate": 0.00016407208668704895, + "loss": 0.2816, + "step": 34550 + }, + { + "epoch": 0.5390902851438198, + "grad_norm": 1.7459170818328857, + "learning_rate": 0.00016406168756889417, + "loss": 0.0834, + "step": 34560 + }, + { + "epoch": 0.5392462719161415, + "grad_norm": 0.89351487159729, + "learning_rate": 0.0001640512884507394, + "loss": 0.1146, + "step": 34570 + }, + { + "epoch": 0.5394022586884633, + "grad_norm": 0.49410831928253174, + "learning_rate": 0.0001640408893325846, + "loss": 0.2319, + "step": 34580 + }, + { + "epoch": 0.5395582454607849, + "grad_norm": 0.009971237741410732, + "learning_rate": 0.00016403049021442983, + "loss": 0.1674, + "step": 34590 + }, + { + "epoch": 0.5397142322331067, + "grad_norm": 1.420511245727539, + "learning_rate": 0.00016402009109627505, + "loss": 0.1469, + "step": 34600 + }, + { + "epoch": 0.5398702190054283, + "grad_norm": 2.5686769485473633, + "learning_rate": 0.00016400969197812027, + "loss": 0.2607, + "step": 34610 + }, + { + "epoch": 0.5400262057777501, + "grad_norm": 2.5808913707733154, + "learning_rate": 0.00016399929285996549, + "loss": 0.4412, + "step": 34620 + }, + { + "epoch": 0.5401821925500717, + "grad_norm": 0.2778591811656952, + "learning_rate": 0.0001639888937418107, + "loss": 0.1563, + "step": 34630 + }, + { + "epoch": 0.5403381793223935, + "grad_norm": 1.6448099613189697, + "learning_rate": 0.00016397849462365592, + "loss": 0.2484, + "step": 34640 + }, + { + "epoch": 0.5404941660947151, + "grad_norm": 0.003213417250663042, + "learning_rate": 0.00016396809550550114, + "loss": 0.1582, + "step": 34650 + }, + { + "epoch": 0.5406501528670369, + "grad_norm": 2.3433990478515625, + "learning_rate": 0.00016395769638734636, + "loss": 0.2774, + "step": 34660 + }, + { + "epoch": 0.5408061396393585, + "grad_norm": 0.09544426947832108, + "learning_rate": 0.00016394729726919158, + "loss": 0.1242, + "step": 34670 + }, + { + "epoch": 0.5409621264116803, + "grad_norm": 1.8934235572814941, + "learning_rate": 0.0001639368981510368, + "loss": 0.1636, + "step": 34680 + }, + { + "epoch": 0.541118113184002, + "grad_norm": 2.8528554439544678, + "learning_rate": 0.00016392649903288202, + "loss": 0.2524, + "step": 34690 + }, + { + "epoch": 0.5412740999563237, + "grad_norm": 2.837836980819702, + "learning_rate": 0.00016391609991472722, + "loss": 0.1808, + "step": 34700 + }, + { + "epoch": 0.5414300867286455, + "grad_norm": 4.349740028381348, + "learning_rate": 0.00016390570079657246, + "loss": 0.12, + "step": 34710 + }, + { + "epoch": 0.5415860735009671, + "grad_norm": 0.6011945605278015, + "learning_rate": 0.00016389530167841765, + "loss": 0.2865, + "step": 34720 + }, + { + "epoch": 0.5417420602732889, + "grad_norm": 1.2079488039016724, + "learning_rate": 0.0001638849025602629, + "loss": 0.125, + "step": 34730 + }, + { + "epoch": 0.5418980470456105, + "grad_norm": 1.9549164772033691, + "learning_rate": 0.0001638745034421081, + "loss": 0.3285, + "step": 34740 + }, + { + "epoch": 0.5420540338179323, + "grad_norm": 0.6969407200813293, + "learning_rate": 0.00016386410432395334, + "loss": 0.3267, + "step": 34750 + }, + { + "epoch": 0.5422100205902539, + "grad_norm": 1.0561965703964233, + "learning_rate": 0.00016385370520579853, + "loss": 0.1463, + "step": 34760 + }, + { + "epoch": 0.5423660073625757, + "grad_norm": 1.8208931684494019, + "learning_rate": 0.00016384330608764378, + "loss": 0.1713, + "step": 34770 + }, + { + "epoch": 0.5425219941348973, + "grad_norm": 0.6349910497665405, + "learning_rate": 0.000163832906969489, + "loss": 0.2244, + "step": 34780 + }, + { + "epoch": 0.5426779809072191, + "grad_norm": 1.1976515054702759, + "learning_rate": 0.00016382250785133422, + "loss": 0.3407, + "step": 34790 + }, + { + "epoch": 0.5428339676795407, + "grad_norm": 0.5970319509506226, + "learning_rate": 0.00016381210873317944, + "loss": 0.1419, + "step": 34800 + }, + { + "epoch": 0.5429899544518625, + "grad_norm": 0.14623159170150757, + "learning_rate": 0.00016380170961502466, + "loss": 0.14, + "step": 34810 + }, + { + "epoch": 0.5431459412241841, + "grad_norm": 1.4332351684570312, + "learning_rate": 0.00016379131049686988, + "loss": 0.1093, + "step": 34820 + }, + { + "epoch": 0.5433019279965059, + "grad_norm": 1.1443063020706177, + "learning_rate": 0.0001637809113787151, + "loss": 0.3676, + "step": 34830 + }, + { + "epoch": 0.5434579147688277, + "grad_norm": 1.3897461891174316, + "learning_rate": 0.00016377051226056032, + "loss": 0.3713, + "step": 34840 + }, + { + "epoch": 0.5436139015411493, + "grad_norm": 1.7773199081420898, + "learning_rate": 0.00016376011314240554, + "loss": 0.2249, + "step": 34850 + }, + { + "epoch": 0.5437698883134711, + "grad_norm": 0.8524389266967773, + "learning_rate": 0.00016374971402425076, + "loss": 0.2158, + "step": 34860 + }, + { + "epoch": 0.5439258750857927, + "grad_norm": 0.5572389960289001, + "learning_rate": 0.00016373931490609598, + "loss": 0.215, + "step": 34870 + }, + { + "epoch": 0.5440818618581145, + "grad_norm": 3.0238709449768066, + "learning_rate": 0.0001637289157879412, + "loss": 0.3765, + "step": 34880 + }, + { + "epoch": 0.5442378486304361, + "grad_norm": 1.1655421257019043, + "learning_rate": 0.00016371851666978642, + "loss": 0.2491, + "step": 34890 + }, + { + "epoch": 0.5443938354027579, + "grad_norm": 0.49492090940475464, + "learning_rate": 0.00016370811755163164, + "loss": 0.1287, + "step": 34900 + }, + { + "epoch": 0.5445498221750795, + "grad_norm": 0.2732921540737152, + "learning_rate": 0.00016369771843347685, + "loss": 0.2154, + "step": 34910 + }, + { + "epoch": 0.5447058089474013, + "grad_norm": 2.5807769298553467, + "learning_rate": 0.00016368731931532207, + "loss": 0.2032, + "step": 34920 + }, + { + "epoch": 0.5448617957197229, + "grad_norm": 0.5044315457344055, + "learning_rate": 0.0001636769201971673, + "loss": 0.2595, + "step": 34930 + }, + { + "epoch": 0.5450177824920447, + "grad_norm": 2.1248972415924072, + "learning_rate": 0.00016366652107901251, + "loss": 0.5651, + "step": 34940 + }, + { + "epoch": 0.5451737692643663, + "grad_norm": 0.8391468524932861, + "learning_rate": 0.00016365612196085773, + "loss": 0.3028, + "step": 34950 + }, + { + "epoch": 0.5453297560366881, + "grad_norm": 0.36081477999687195, + "learning_rate": 0.00016364572284270295, + "loss": 0.1073, + "step": 34960 + }, + { + "epoch": 0.5454857428090097, + "grad_norm": 3.175804376602173, + "learning_rate": 0.00016363532372454817, + "loss": 0.1889, + "step": 34970 + }, + { + "epoch": 0.5456417295813315, + "grad_norm": 0.20647937059402466, + "learning_rate": 0.0001636249246063934, + "loss": 0.2416, + "step": 34980 + }, + { + "epoch": 0.5457977163536533, + "grad_norm": 2.3525943756103516, + "learning_rate": 0.0001636145254882386, + "loss": 0.3179, + "step": 34990 + }, + { + "epoch": 0.5459537031259749, + "grad_norm": 0.2641216814517975, + "learning_rate": 0.00016360412637008383, + "loss": 0.2157, + "step": 35000 + }, + { + "epoch": 0.5461096898982967, + "grad_norm": 0.09906073659658432, + "learning_rate": 0.00016359372725192905, + "loss": 0.2576, + "step": 35010 + }, + { + "epoch": 0.5462656766706183, + "grad_norm": 0.13971920311450958, + "learning_rate": 0.00016358332813377427, + "loss": 0.1298, + "step": 35020 + }, + { + "epoch": 0.5464216634429401, + "grad_norm": 1.4535638093948364, + "learning_rate": 0.0001635729290156195, + "loss": 0.2357, + "step": 35030 + }, + { + "epoch": 0.5465776502152617, + "grad_norm": 0.9107828736305237, + "learning_rate": 0.0001635625298974647, + "loss": 0.3049, + "step": 35040 + }, + { + "epoch": 0.5467336369875835, + "grad_norm": 0.0377386212348938, + "learning_rate": 0.00016355213077930993, + "loss": 0.2066, + "step": 35050 + }, + { + "epoch": 0.5468896237599051, + "grad_norm": 2.7800869941711426, + "learning_rate": 0.00016354173166115515, + "loss": 0.2974, + "step": 35060 + }, + { + "epoch": 0.5470456105322269, + "grad_norm": 1.6247998476028442, + "learning_rate": 0.00016353133254300037, + "loss": 0.3822, + "step": 35070 + }, + { + "epoch": 0.5472015973045485, + "grad_norm": 1.6479015350341797, + "learning_rate": 0.0001635209334248456, + "loss": 0.3401, + "step": 35080 + }, + { + "epoch": 0.5473575840768703, + "grad_norm": 0.1879737675189972, + "learning_rate": 0.0001635105343066908, + "loss": 0.1549, + "step": 35090 + }, + { + "epoch": 0.5475135708491919, + "grad_norm": 0.39355361461639404, + "learning_rate": 0.00016350013518853603, + "loss": 0.2032, + "step": 35100 + }, + { + "epoch": 0.5476695576215137, + "grad_norm": 2.9976983070373535, + "learning_rate": 0.00016348973607038125, + "loss": 0.3867, + "step": 35110 + }, + { + "epoch": 0.5478255443938354, + "grad_norm": 0.8242707252502441, + "learning_rate": 0.00016347933695222647, + "loss": 0.3601, + "step": 35120 + }, + { + "epoch": 0.5479815311661571, + "grad_norm": 0.874955415725708, + "learning_rate": 0.0001634689378340717, + "loss": 0.2709, + "step": 35130 + }, + { + "epoch": 0.5481375179384789, + "grad_norm": 0.05980971083045006, + "learning_rate": 0.0001634585387159169, + "loss": 0.2646, + "step": 35140 + }, + { + "epoch": 0.5482935047108005, + "grad_norm": 1.8437328338623047, + "learning_rate": 0.0001634481395977621, + "loss": 0.1965, + "step": 35150 + }, + { + "epoch": 0.5484494914831223, + "grad_norm": 0.4142405688762665, + "learning_rate": 0.00016343774047960735, + "loss": 0.1756, + "step": 35160 + }, + { + "epoch": 0.5486054782554439, + "grad_norm": 4.836324214935303, + "learning_rate": 0.00016342734136145254, + "loss": 0.2534, + "step": 35170 + }, + { + "epoch": 0.5487614650277657, + "grad_norm": 1.4933065176010132, + "learning_rate": 0.00016341694224329779, + "loss": 0.2355, + "step": 35180 + }, + { + "epoch": 0.5489174518000873, + "grad_norm": 1.9468894004821777, + "learning_rate": 0.00016340654312514298, + "loss": 0.3113, + "step": 35190 + }, + { + "epoch": 0.5490734385724091, + "grad_norm": 1.012710452079773, + "learning_rate": 0.00016339614400698822, + "loss": 0.2635, + "step": 35200 + }, + { + "epoch": 0.5492294253447307, + "grad_norm": 1.5551140308380127, + "learning_rate": 0.00016338574488883342, + "loss": 0.2951, + "step": 35210 + }, + { + "epoch": 0.5493854121170525, + "grad_norm": 1.8613696098327637, + "learning_rate": 0.00016337534577067866, + "loss": 0.2832, + "step": 35220 + }, + { + "epoch": 0.5495413988893741, + "grad_norm": 1.6748839616775513, + "learning_rate": 0.00016336494665252386, + "loss": 0.2266, + "step": 35230 + }, + { + "epoch": 0.5496973856616959, + "grad_norm": 0.4094032347202301, + "learning_rate": 0.0001633545475343691, + "loss": 0.3497, + "step": 35240 + }, + { + "epoch": 0.5498533724340176, + "grad_norm": 1.6414631605148315, + "learning_rate": 0.0001633441484162143, + "loss": 0.1763, + "step": 35250 + }, + { + "epoch": 0.5500093592063393, + "grad_norm": 1.6380645036697388, + "learning_rate": 0.00016333374929805954, + "loss": 0.1355, + "step": 35260 + }, + { + "epoch": 0.550165345978661, + "grad_norm": 0.7325630187988281, + "learning_rate": 0.00016332335017990473, + "loss": 0.3105, + "step": 35270 + }, + { + "epoch": 0.5503213327509827, + "grad_norm": 1.6015644073486328, + "learning_rate": 0.00016331295106174998, + "loss": 0.2311, + "step": 35280 + }, + { + "epoch": 0.5504773195233045, + "grad_norm": 0.9224210381507874, + "learning_rate": 0.00016330255194359517, + "loss": 0.2746, + "step": 35290 + }, + { + "epoch": 0.5506333062956261, + "grad_norm": 1.30025315284729, + "learning_rate": 0.00016329215282544042, + "loss": 0.2066, + "step": 35300 + }, + { + "epoch": 0.5507892930679479, + "grad_norm": 0.481125146150589, + "learning_rate": 0.0001632817537072856, + "loss": 0.1778, + "step": 35310 + }, + { + "epoch": 0.5509452798402695, + "grad_norm": 0.2709486782550812, + "learning_rate": 0.00016327135458913086, + "loss": 0.1131, + "step": 35320 + }, + { + "epoch": 0.5511012666125913, + "grad_norm": 0.6248563528060913, + "learning_rate": 0.00016326095547097605, + "loss": 0.4356, + "step": 35330 + }, + { + "epoch": 0.5512572533849129, + "grad_norm": 3.0947647094726562, + "learning_rate": 0.0001632505563528213, + "loss": 0.3072, + "step": 35340 + }, + { + "epoch": 0.5514132401572347, + "grad_norm": 1.009535312652588, + "learning_rate": 0.0001632401572346665, + "loss": 0.1808, + "step": 35350 + }, + { + "epoch": 0.5515692269295563, + "grad_norm": 0.042604975402355194, + "learning_rate": 0.00016322975811651174, + "loss": 0.1198, + "step": 35360 + }, + { + "epoch": 0.5517252137018781, + "grad_norm": 1.8652396202087402, + "learning_rate": 0.00016321935899835693, + "loss": 0.2176, + "step": 35370 + }, + { + "epoch": 0.5518812004741998, + "grad_norm": 4.6097187995910645, + "learning_rate": 0.00016320895988020218, + "loss": 0.2132, + "step": 35380 + }, + { + "epoch": 0.5520371872465215, + "grad_norm": 1.3615522384643555, + "learning_rate": 0.00016319856076204737, + "loss": 0.2751, + "step": 35390 + }, + { + "epoch": 0.5521931740188432, + "grad_norm": 0.8844773769378662, + "learning_rate": 0.00016318816164389262, + "loss": 0.22, + "step": 35400 + }, + { + "epoch": 0.5523491607911649, + "grad_norm": 0.5413331985473633, + "learning_rate": 0.0001631777625257378, + "loss": 0.2159, + "step": 35410 + }, + { + "epoch": 0.5525051475634866, + "grad_norm": 1.5456678867340088, + "learning_rate": 0.00016316736340758306, + "loss": 0.2244, + "step": 35420 + }, + { + "epoch": 0.5526611343358083, + "grad_norm": 2.1405861377716064, + "learning_rate": 0.00016315696428942825, + "loss": 0.446, + "step": 35430 + }, + { + "epoch": 0.5528171211081301, + "grad_norm": 1.4269858598709106, + "learning_rate": 0.0001631465651712735, + "loss": 0.3599, + "step": 35440 + }, + { + "epoch": 0.5529731078804517, + "grad_norm": 1.8682516813278198, + "learning_rate": 0.0001631361660531187, + "loss": 0.115, + "step": 35450 + }, + { + "epoch": 0.5531290946527735, + "grad_norm": 1.0175774097442627, + "learning_rate": 0.00016312576693496394, + "loss": 0.4003, + "step": 35460 + }, + { + "epoch": 0.5532850814250951, + "grad_norm": 2.4369170665740967, + "learning_rate": 0.00016311536781680913, + "loss": 0.2489, + "step": 35470 + }, + { + "epoch": 0.5534410681974169, + "grad_norm": 1.886022686958313, + "learning_rate": 0.00016310496869865437, + "loss": 0.4201, + "step": 35480 + }, + { + "epoch": 0.5535970549697385, + "grad_norm": 1.5811959505081177, + "learning_rate": 0.00016309456958049957, + "loss": 0.4286, + "step": 35490 + }, + { + "epoch": 0.5537530417420603, + "grad_norm": 1.971110224723816, + "learning_rate": 0.00016308417046234481, + "loss": 0.3599, + "step": 35500 + }, + { + "epoch": 0.553909028514382, + "grad_norm": 0.0335380844771862, + "learning_rate": 0.00016307377134419, + "loss": 0.3357, + "step": 35510 + }, + { + "epoch": 0.5540650152867037, + "grad_norm": 1.8519577980041504, + "learning_rate": 0.00016306337222603525, + "loss": 0.3126, + "step": 35520 + }, + { + "epoch": 0.5542210020590254, + "grad_norm": 2.1563379764556885, + "learning_rate": 0.00016305297310788045, + "loss": 0.1123, + "step": 35530 + }, + { + "epoch": 0.5543769888313471, + "grad_norm": 1.64332914352417, + "learning_rate": 0.0001630425739897257, + "loss": 0.2853, + "step": 35540 + }, + { + "epoch": 0.5545329756036688, + "grad_norm": 0.061150554567575455, + "learning_rate": 0.00016303217487157088, + "loss": 0.1943, + "step": 35550 + }, + { + "epoch": 0.5546889623759905, + "grad_norm": 1.2701060771942139, + "learning_rate": 0.00016302177575341613, + "loss": 0.2304, + "step": 35560 + }, + { + "epoch": 0.5548449491483122, + "grad_norm": 2.424860715866089, + "learning_rate": 0.00016301137663526132, + "loss": 0.2128, + "step": 35570 + }, + { + "epoch": 0.5550009359206339, + "grad_norm": 0.6803575158119202, + "learning_rate": 0.00016300097751710657, + "loss": 0.1145, + "step": 35580 + }, + { + "epoch": 0.5551569226929557, + "grad_norm": 1.2855092287063599, + "learning_rate": 0.00016299057839895176, + "loss": 0.2448, + "step": 35590 + }, + { + "epoch": 0.5553129094652773, + "grad_norm": 2.6340911388397217, + "learning_rate": 0.00016298017928079698, + "loss": 0.1856, + "step": 35600 + }, + { + "epoch": 0.5554688962375991, + "grad_norm": 1.776382327079773, + "learning_rate": 0.0001629697801626422, + "loss": 0.2422, + "step": 35610 + }, + { + "epoch": 0.5556248830099207, + "grad_norm": 3.0746347904205322, + "learning_rate": 0.00016295938104448742, + "loss": 0.1578, + "step": 35620 + }, + { + "epoch": 0.5557808697822425, + "grad_norm": 1.4887659549713135, + "learning_rate": 0.00016294898192633267, + "loss": 0.2385, + "step": 35630 + }, + { + "epoch": 0.5559368565545642, + "grad_norm": 0.5371220707893372, + "learning_rate": 0.00016293858280817786, + "loss": 0.2361, + "step": 35640 + }, + { + "epoch": 0.5560928433268859, + "grad_norm": 0.18604904413223267, + "learning_rate": 0.0001629281836900231, + "loss": 0.0828, + "step": 35650 + }, + { + "epoch": 0.5562488300992076, + "grad_norm": 2.267854690551758, + "learning_rate": 0.0001629177845718683, + "loss": 0.413, + "step": 35660 + }, + { + "epoch": 0.5564048168715293, + "grad_norm": 1.9755452871322632, + "learning_rate": 0.00016290738545371355, + "loss": 0.146, + "step": 35670 + }, + { + "epoch": 0.556560803643851, + "grad_norm": 2.430293321609497, + "learning_rate": 0.00016289698633555874, + "loss": 0.2677, + "step": 35680 + }, + { + "epoch": 0.5567167904161727, + "grad_norm": 0.6319543719291687, + "learning_rate": 0.000162886587217404, + "loss": 0.1481, + "step": 35690 + }, + { + "epoch": 0.5568727771884944, + "grad_norm": 0.17439277470111847, + "learning_rate": 0.00016287618809924918, + "loss": 0.1641, + "step": 35700 + }, + { + "epoch": 0.5570287639608161, + "grad_norm": 0.350175142288208, + "learning_rate": 0.00016286578898109443, + "loss": 0.393, + "step": 35710 + }, + { + "epoch": 0.5571847507331378, + "grad_norm": 2.641941547393799, + "learning_rate": 0.00016285538986293962, + "loss": 0.1907, + "step": 35720 + }, + { + "epoch": 0.5573407375054595, + "grad_norm": 0.1910303384065628, + "learning_rate": 0.00016284499074478487, + "loss": 0.2248, + "step": 35730 + }, + { + "epoch": 0.5574967242777813, + "grad_norm": 1.0180896520614624, + "learning_rate": 0.00016283459162663006, + "loss": 0.1924, + "step": 35740 + }, + { + "epoch": 0.557652711050103, + "grad_norm": 1.2047260999679565, + "learning_rate": 0.0001628241925084753, + "loss": 0.2031, + "step": 35750 + }, + { + "epoch": 0.5578086978224247, + "grad_norm": 1.8182405233383179, + "learning_rate": 0.0001628137933903205, + "loss": 0.1573, + "step": 35760 + }, + { + "epoch": 0.5579646845947464, + "grad_norm": 0.5485963821411133, + "learning_rate": 0.00016280339427216574, + "loss": 0.152, + "step": 35770 + }, + { + "epoch": 0.5581206713670681, + "grad_norm": 0.1674145758152008, + "learning_rate": 0.00016279299515401094, + "loss": 0.2388, + "step": 35780 + }, + { + "epoch": 0.5582766581393898, + "grad_norm": 0.36187657713890076, + "learning_rate": 0.00016278259603585618, + "loss": 0.1618, + "step": 35790 + }, + { + "epoch": 0.5584326449117115, + "grad_norm": 0.013768521137535572, + "learning_rate": 0.00016277219691770138, + "loss": 0.3666, + "step": 35800 + }, + { + "epoch": 0.5585886316840332, + "grad_norm": 2.8602802753448486, + "learning_rate": 0.00016276179779954662, + "loss": 0.2065, + "step": 35810 + }, + { + "epoch": 0.5587446184563549, + "grad_norm": 1.7925455570220947, + "learning_rate": 0.00016275139868139182, + "loss": 0.1729, + "step": 35820 + }, + { + "epoch": 0.5589006052286766, + "grad_norm": 2.910456418991089, + "learning_rate": 0.00016274099956323706, + "loss": 0.3734, + "step": 35830 + }, + { + "epoch": 0.5590565920009983, + "grad_norm": 0.8378308415412903, + "learning_rate": 0.00016273060044508225, + "loss": 0.2255, + "step": 35840 + }, + { + "epoch": 0.55921257877332, + "grad_norm": 0.409534752368927, + "learning_rate": 0.0001627202013269275, + "loss": 0.2046, + "step": 35850 + }, + { + "epoch": 0.5593685655456417, + "grad_norm": 0.49498099088668823, + "learning_rate": 0.0001627098022087727, + "loss": 0.1058, + "step": 35860 + }, + { + "epoch": 0.5595245523179634, + "grad_norm": 1.1761783361434937, + "learning_rate": 0.00016269940309061794, + "loss": 0.1566, + "step": 35870 + }, + { + "epoch": 0.5596805390902851, + "grad_norm": 3.8204751014709473, + "learning_rate": 0.00016268900397246313, + "loss": 0.2646, + "step": 35880 + }, + { + "epoch": 0.5598365258626069, + "grad_norm": 0.9882522225379944, + "learning_rate": 0.00016267860485430838, + "loss": 0.1756, + "step": 35890 + }, + { + "epoch": 0.5599925126349286, + "grad_norm": 1.1832259893417358, + "learning_rate": 0.00016266820573615357, + "loss": 0.1385, + "step": 35900 + }, + { + "epoch": 0.5601484994072503, + "grad_norm": 0.7638296484947205, + "learning_rate": 0.00016265780661799882, + "loss": 0.5679, + "step": 35910 + }, + { + "epoch": 0.560304486179572, + "grad_norm": 0.8551504611968994, + "learning_rate": 0.000162647407499844, + "loss": 0.2981, + "step": 35920 + }, + { + "epoch": 0.5604604729518937, + "grad_norm": 5.499948501586914, + "learning_rate": 0.00016263700838168926, + "loss": 0.1429, + "step": 35930 + }, + { + "epoch": 0.5606164597242154, + "grad_norm": 3.0494496822357178, + "learning_rate": 0.00016262660926353445, + "loss": 0.1322, + "step": 35940 + }, + { + "epoch": 0.5607724464965371, + "grad_norm": 0.2482384592294693, + "learning_rate": 0.0001626162101453797, + "loss": 0.2473, + "step": 35950 + }, + { + "epoch": 0.5609284332688588, + "grad_norm": 0.009750776924192905, + "learning_rate": 0.0001626058110272249, + "loss": 0.0429, + "step": 35960 + }, + { + "epoch": 0.5610844200411805, + "grad_norm": 0.5742604732513428, + "learning_rate": 0.00016259541190907014, + "loss": 0.4565, + "step": 35970 + }, + { + "epoch": 0.5612404068135022, + "grad_norm": 0.7810243368148804, + "learning_rate": 0.00016258501279091533, + "loss": 0.2107, + "step": 35980 + }, + { + "epoch": 0.5613963935858239, + "grad_norm": 2.785747528076172, + "learning_rate": 0.00016257461367276058, + "loss": 0.5087, + "step": 35990 + }, + { + "epoch": 0.5615523803581456, + "grad_norm": 0.8806902766227722, + "learning_rate": 0.00016256421455460577, + "loss": 0.2937, + "step": 36000 + }, + { + "epoch": 0.5617083671304673, + "grad_norm": 1.857373833656311, + "learning_rate": 0.00016255381543645102, + "loss": 0.5042, + "step": 36010 + }, + { + "epoch": 0.561864353902789, + "grad_norm": 0.5333901047706604, + "learning_rate": 0.0001625434163182962, + "loss": 0.1678, + "step": 36020 + }, + { + "epoch": 0.5620203406751108, + "grad_norm": 11.023160934448242, + "learning_rate": 0.00016253301720014145, + "loss": 0.2721, + "step": 36030 + }, + { + "epoch": 0.5621763274474325, + "grad_norm": 2.6214029788970947, + "learning_rate": 0.00016252261808198665, + "loss": 0.1675, + "step": 36040 + }, + { + "epoch": 0.5623323142197542, + "grad_norm": 0.25980064272880554, + "learning_rate": 0.0001625122189638319, + "loss": 0.1832, + "step": 36050 + }, + { + "epoch": 0.5624883009920759, + "grad_norm": 1.3559473752975464, + "learning_rate": 0.0001625018198456771, + "loss": 0.121, + "step": 36060 + }, + { + "epoch": 0.5626442877643976, + "grad_norm": 0.3558153212070465, + "learning_rate": 0.0001624914207275223, + "loss": 0.2797, + "step": 36070 + }, + { + "epoch": 0.5628002745367193, + "grad_norm": 0.6314427256584167, + "learning_rate": 0.00016248102160936753, + "loss": 0.3455, + "step": 36080 + }, + { + "epoch": 0.562956261309041, + "grad_norm": 0.16120101511478424, + "learning_rate": 0.00016247062249121275, + "loss": 0.3669, + "step": 36090 + }, + { + "epoch": 0.5631122480813627, + "grad_norm": 1.041572093963623, + "learning_rate": 0.00016246022337305797, + "loss": 0.2744, + "step": 36100 + }, + { + "epoch": 0.5632682348536844, + "grad_norm": 0.2645890414714813, + "learning_rate": 0.00016244982425490318, + "loss": 0.1487, + "step": 36110 + }, + { + "epoch": 0.5634242216260061, + "grad_norm": 5.698698043823242, + "learning_rate": 0.0001624394251367484, + "loss": 0.173, + "step": 36120 + }, + { + "epoch": 0.5635802083983278, + "grad_norm": 0.3299804627895355, + "learning_rate": 0.00016242902601859362, + "loss": 0.1965, + "step": 36130 + }, + { + "epoch": 0.5637361951706495, + "grad_norm": 1.2476481199264526, + "learning_rate": 0.00016241862690043884, + "loss": 0.4097, + "step": 36140 + }, + { + "epoch": 0.5638921819429712, + "grad_norm": 0.30926263332366943, + "learning_rate": 0.00016240822778228406, + "loss": 0.1416, + "step": 36150 + }, + { + "epoch": 0.564048168715293, + "grad_norm": 0.05872740224003792, + "learning_rate": 0.00016239782866412928, + "loss": 0.1449, + "step": 36160 + }, + { + "epoch": 0.5642041554876146, + "grad_norm": 0.2292211800813675, + "learning_rate": 0.0001623874295459745, + "loss": 0.2557, + "step": 36170 + }, + { + "epoch": 0.5643601422599364, + "grad_norm": 1.7822531461715698, + "learning_rate": 0.00016237703042781972, + "loss": 0.2953, + "step": 36180 + }, + { + "epoch": 0.5645161290322581, + "grad_norm": 2.4908461570739746, + "learning_rate": 0.00016236663130966494, + "loss": 0.3888, + "step": 36190 + }, + { + "epoch": 0.5646721158045798, + "grad_norm": 0.08363594114780426, + "learning_rate": 0.00016235623219151016, + "loss": 0.2293, + "step": 36200 + }, + { + "epoch": 0.5648281025769015, + "grad_norm": 1.4830002784729004, + "learning_rate": 0.00016234583307335538, + "loss": 0.2348, + "step": 36210 + }, + { + "epoch": 0.5649840893492232, + "grad_norm": 1.8443071842193604, + "learning_rate": 0.0001623354339552006, + "loss": 0.3392, + "step": 36220 + }, + { + "epoch": 0.5651400761215449, + "grad_norm": 2.8050875663757324, + "learning_rate": 0.00016232503483704582, + "loss": 0.2426, + "step": 36230 + }, + { + "epoch": 0.5652960628938666, + "grad_norm": 3.6627259254455566, + "learning_rate": 0.00016231463571889104, + "loss": 0.2118, + "step": 36240 + }, + { + "epoch": 0.5654520496661883, + "grad_norm": 0.3443094789981842, + "learning_rate": 0.00016230423660073626, + "loss": 0.2418, + "step": 36250 + }, + { + "epoch": 0.56560803643851, + "grad_norm": 1.7411112785339355, + "learning_rate": 0.00016229383748258148, + "loss": 0.1628, + "step": 36260 + }, + { + "epoch": 0.5657640232108317, + "grad_norm": 1.1321816444396973, + "learning_rate": 0.0001622834383644267, + "loss": 0.2732, + "step": 36270 + }, + { + "epoch": 0.5659200099831534, + "grad_norm": 2.9686615467071533, + "learning_rate": 0.00016227303924627192, + "loss": 0.3036, + "step": 36280 + }, + { + "epoch": 0.5660759967554752, + "grad_norm": 5.03530740737915, + "learning_rate": 0.00016226264012811714, + "loss": 0.2143, + "step": 36290 + }, + { + "epoch": 0.5662319835277968, + "grad_norm": 2.3943281173706055, + "learning_rate": 0.00016225224100996236, + "loss": 0.2629, + "step": 36300 + }, + { + "epoch": 0.5663879703001186, + "grad_norm": 1.812828779220581, + "learning_rate": 0.00016224184189180758, + "loss": 0.3892, + "step": 36310 + }, + { + "epoch": 0.5665439570724402, + "grad_norm": 1.4708483219146729, + "learning_rate": 0.0001622314427736528, + "loss": 0.2118, + "step": 36320 + }, + { + "epoch": 0.566699943844762, + "grad_norm": 8.99713134765625, + "learning_rate": 0.00016222104365549802, + "loss": 0.3185, + "step": 36330 + }, + { + "epoch": 0.5668559306170837, + "grad_norm": 1.7472341060638428, + "learning_rate": 0.00016221064453734324, + "loss": 0.1659, + "step": 36340 + }, + { + "epoch": 0.5670119173894054, + "grad_norm": 0.8059778213500977, + "learning_rate": 0.00016220024541918846, + "loss": 0.0621, + "step": 36350 + }, + { + "epoch": 0.5671679041617271, + "grad_norm": 4.0501861572265625, + "learning_rate": 0.00016218984630103368, + "loss": 0.1399, + "step": 36360 + }, + { + "epoch": 0.5673238909340488, + "grad_norm": 0.014400321058928967, + "learning_rate": 0.0001621794471828789, + "loss": 0.2372, + "step": 36370 + }, + { + "epoch": 0.5674798777063705, + "grad_norm": 0.4061933755874634, + "learning_rate": 0.00016216904806472412, + "loss": 0.3854, + "step": 36380 + }, + { + "epoch": 0.5676358644786922, + "grad_norm": 0.4940955340862274, + "learning_rate": 0.00016215864894656933, + "loss": 0.1689, + "step": 36390 + }, + { + "epoch": 0.567791851251014, + "grad_norm": 0.3220517635345459, + "learning_rate": 0.00016214824982841455, + "loss": 0.1912, + "step": 36400 + }, + { + "epoch": 0.5679478380233356, + "grad_norm": 2.0848920345306396, + "learning_rate": 0.00016213785071025977, + "loss": 0.1639, + "step": 36410 + }, + { + "epoch": 0.5681038247956574, + "grad_norm": 0.04129406809806824, + "learning_rate": 0.000162127451592105, + "loss": 0.1378, + "step": 36420 + }, + { + "epoch": 0.568259811567979, + "grad_norm": 1.9343401193618774, + "learning_rate": 0.0001621170524739502, + "loss": 0.3759, + "step": 36430 + }, + { + "epoch": 0.5684157983403008, + "grad_norm": 1.0625615119934082, + "learning_rate": 0.00016210665335579543, + "loss": 0.2074, + "step": 36440 + }, + { + "epoch": 0.5685717851126224, + "grad_norm": 0.14784491062164307, + "learning_rate": 0.00016209625423764065, + "loss": 0.0981, + "step": 36450 + }, + { + "epoch": 0.5687277718849442, + "grad_norm": 1.8117821216583252, + "learning_rate": 0.00016208585511948587, + "loss": 0.2755, + "step": 36460 + }, + { + "epoch": 0.5688837586572658, + "grad_norm": 0.061010006815195084, + "learning_rate": 0.0001620754560013311, + "loss": 0.1815, + "step": 36470 + }, + { + "epoch": 0.5690397454295876, + "grad_norm": 0.44453224539756775, + "learning_rate": 0.0001620650568831763, + "loss": 0.0891, + "step": 36480 + }, + { + "epoch": 0.5691957322019093, + "grad_norm": 0.6130645871162415, + "learning_rate": 0.00016205465776502153, + "loss": 0.2778, + "step": 36490 + }, + { + "epoch": 0.569351718974231, + "grad_norm": 5.408796787261963, + "learning_rate": 0.00016204425864686678, + "loss": 0.2301, + "step": 36500 + }, + { + "epoch": 0.5695077057465527, + "grad_norm": 1.2095412015914917, + "learning_rate": 0.00016203385952871197, + "loss": 0.1378, + "step": 36510 + }, + { + "epoch": 0.5696636925188744, + "grad_norm": 0.46873530745506287, + "learning_rate": 0.0001620234604105572, + "loss": 0.1431, + "step": 36520 + }, + { + "epoch": 0.5698196792911961, + "grad_norm": 1.028438925743103, + "learning_rate": 0.0001620130612924024, + "loss": 0.1114, + "step": 36530 + }, + { + "epoch": 0.5699756660635178, + "grad_norm": 0.5952587127685547, + "learning_rate": 0.00016200266217424763, + "loss": 0.2156, + "step": 36540 + }, + { + "epoch": 0.5701316528358396, + "grad_norm": 0.5134285688400269, + "learning_rate": 0.00016199226305609285, + "loss": 0.2712, + "step": 36550 + }, + { + "epoch": 0.5702876396081612, + "grad_norm": 0.03440163657069206, + "learning_rate": 0.00016198186393793807, + "loss": 0.312, + "step": 36560 + }, + { + "epoch": 0.570443626380483, + "grad_norm": 0.10363951325416565, + "learning_rate": 0.0001619714648197833, + "loss": 0.2399, + "step": 36570 + }, + { + "epoch": 0.5705996131528046, + "grad_norm": 1.3341947793960571, + "learning_rate": 0.0001619610657016285, + "loss": 0.1397, + "step": 36580 + }, + { + "epoch": 0.5707555999251264, + "grad_norm": 0.0668073296546936, + "learning_rate": 0.00016195066658347373, + "loss": 0.1091, + "step": 36590 + }, + { + "epoch": 0.570911586697448, + "grad_norm": 5.0486674308776855, + "learning_rate": 0.00016194026746531895, + "loss": 0.4893, + "step": 36600 + }, + { + "epoch": 0.5710675734697698, + "grad_norm": 1.2655692100524902, + "learning_rate": 0.00016192986834716417, + "loss": 0.3344, + "step": 36610 + }, + { + "epoch": 0.5712235602420914, + "grad_norm": 0.1458190381526947, + "learning_rate": 0.00016191946922900939, + "loss": 0.247, + "step": 36620 + }, + { + "epoch": 0.5713795470144132, + "grad_norm": 2.5789072513580322, + "learning_rate": 0.0001619090701108546, + "loss": 0.1546, + "step": 36630 + }, + { + "epoch": 0.5715355337867349, + "grad_norm": 1.191821575164795, + "learning_rate": 0.00016189867099269983, + "loss": 0.2215, + "step": 36640 + }, + { + "epoch": 0.5716915205590566, + "grad_norm": 8.163617134094238, + "learning_rate": 0.00016188827187454505, + "loss": 0.1696, + "step": 36650 + }, + { + "epoch": 0.5718475073313783, + "grad_norm": 3.162277936935425, + "learning_rate": 0.00016187787275639027, + "loss": 0.2906, + "step": 36660 + }, + { + "epoch": 0.5720034941037, + "grad_norm": 2.2368602752685547, + "learning_rate": 0.00016186747363823548, + "loss": 0.3806, + "step": 36670 + }, + { + "epoch": 0.5721594808760218, + "grad_norm": 2.0193803310394287, + "learning_rate": 0.0001618570745200807, + "loss": 0.1964, + "step": 36680 + }, + { + "epoch": 0.5723154676483434, + "grad_norm": 0.9131811261177063, + "learning_rate": 0.00016184667540192592, + "loss": 0.1902, + "step": 36690 + }, + { + "epoch": 0.5724714544206652, + "grad_norm": 0.21330870687961578, + "learning_rate": 0.00016183627628377114, + "loss": 0.2907, + "step": 36700 + }, + { + "epoch": 0.5726274411929868, + "grad_norm": 0.7145895957946777, + "learning_rate": 0.00016182587716561636, + "loss": 0.2574, + "step": 36710 + }, + { + "epoch": 0.5727834279653086, + "grad_norm": 1.0699433088302612, + "learning_rate": 0.00016181547804746158, + "loss": 0.1675, + "step": 36720 + }, + { + "epoch": 0.5729394147376302, + "grad_norm": 0.5066679120063782, + "learning_rate": 0.0001618050789293068, + "loss": 0.4487, + "step": 36730 + }, + { + "epoch": 0.573095401509952, + "grad_norm": 0.7482333183288574, + "learning_rate": 0.00016179467981115202, + "loss": 0.4101, + "step": 36740 + }, + { + "epoch": 0.5732513882822736, + "grad_norm": 0.15925532579421997, + "learning_rate": 0.00016178428069299724, + "loss": 0.2825, + "step": 36750 + }, + { + "epoch": 0.5734073750545954, + "grad_norm": 0.95689457654953, + "learning_rate": 0.00016177388157484246, + "loss": 0.2754, + "step": 36760 + }, + { + "epoch": 0.573563361826917, + "grad_norm": 2.15043568611145, + "learning_rate": 0.00016176348245668768, + "loss": 0.2073, + "step": 36770 + }, + { + "epoch": 0.5737193485992388, + "grad_norm": 2.651520252227783, + "learning_rate": 0.0001617530833385329, + "loss": 0.2517, + "step": 36780 + }, + { + "epoch": 0.5738753353715605, + "grad_norm": 2.1156702041625977, + "learning_rate": 0.00016174268422037812, + "loss": 0.3029, + "step": 36790 + }, + { + "epoch": 0.5740313221438822, + "grad_norm": 3.6200709342956543, + "learning_rate": 0.00016173228510222334, + "loss": 0.4174, + "step": 36800 + }, + { + "epoch": 0.574187308916204, + "grad_norm": 1.447937250137329, + "learning_rate": 0.00016172188598406856, + "loss": 0.1763, + "step": 36810 + }, + { + "epoch": 0.5743432956885256, + "grad_norm": 1.4020819664001465, + "learning_rate": 0.00016171148686591378, + "loss": 0.3994, + "step": 36820 + }, + { + "epoch": 0.5744992824608474, + "grad_norm": 1.4079807996749878, + "learning_rate": 0.000161701087747759, + "loss": 0.2051, + "step": 36830 + }, + { + "epoch": 0.574655269233169, + "grad_norm": 0.05268567427992821, + "learning_rate": 0.00016169068862960422, + "loss": 0.1773, + "step": 36840 + }, + { + "epoch": 0.5748112560054908, + "grad_norm": 2.248767614364624, + "learning_rate": 0.00016168028951144944, + "loss": 0.2534, + "step": 36850 + }, + { + "epoch": 0.5749672427778124, + "grad_norm": 0.3726302981376648, + "learning_rate": 0.00016166989039329466, + "loss": 0.328, + "step": 36860 + }, + { + "epoch": 0.5751232295501342, + "grad_norm": 1.0772110223770142, + "learning_rate": 0.00016165949127513988, + "loss": 0.2069, + "step": 36870 + }, + { + "epoch": 0.5752792163224558, + "grad_norm": 0.11339398473501205, + "learning_rate": 0.0001616490921569851, + "loss": 0.1827, + "step": 36880 + }, + { + "epoch": 0.5754352030947776, + "grad_norm": 1.2981103658676147, + "learning_rate": 0.00016163869303883032, + "loss": 0.1378, + "step": 36890 + }, + { + "epoch": 0.5755911898670992, + "grad_norm": 0.3302285075187683, + "learning_rate": 0.00016162829392067554, + "loss": 0.2311, + "step": 36900 + }, + { + "epoch": 0.575747176639421, + "grad_norm": 0.10168436169624329, + "learning_rate": 0.00016161789480252076, + "loss": 0.1678, + "step": 36910 + }, + { + "epoch": 0.5759031634117426, + "grad_norm": 1.6778357028961182, + "learning_rate": 0.00016160749568436598, + "loss": 0.2385, + "step": 36920 + }, + { + "epoch": 0.5760591501840644, + "grad_norm": 8.1710205078125, + "learning_rate": 0.0001615970965662112, + "loss": 0.3287, + "step": 36930 + }, + { + "epoch": 0.5762151369563862, + "grad_norm": 2.6481053829193115, + "learning_rate": 0.00016158669744805641, + "loss": 0.3119, + "step": 36940 + }, + { + "epoch": 0.5763711237287078, + "grad_norm": 0.3120105266571045, + "learning_rate": 0.00016157629832990163, + "loss": 0.1985, + "step": 36950 + }, + { + "epoch": 0.5765271105010296, + "grad_norm": 2.7221410274505615, + "learning_rate": 0.00016156589921174685, + "loss": 0.3697, + "step": 36960 + }, + { + "epoch": 0.5766830972733512, + "grad_norm": 0.34397706389427185, + "learning_rate": 0.00016155550009359207, + "loss": 0.2162, + "step": 36970 + }, + { + "epoch": 0.576839084045673, + "grad_norm": 4.6421003341674805, + "learning_rate": 0.0001615451009754373, + "loss": 0.2892, + "step": 36980 + }, + { + "epoch": 0.5769950708179946, + "grad_norm": 1.5395855903625488, + "learning_rate": 0.0001615347018572825, + "loss": 0.1258, + "step": 36990 + }, + { + "epoch": 0.5771510575903164, + "grad_norm": 0.4586782455444336, + "learning_rate": 0.00016152430273912773, + "loss": 0.1752, + "step": 37000 + }, + { + "epoch": 0.577307044362638, + "grad_norm": 0.4277658760547638, + "learning_rate": 0.00016151390362097295, + "loss": 0.1722, + "step": 37010 + }, + { + "epoch": 0.5774630311349598, + "grad_norm": 1.0282249450683594, + "learning_rate": 0.00016150350450281817, + "loss": 0.1361, + "step": 37020 + }, + { + "epoch": 0.5776190179072814, + "grad_norm": 0.2496921867132187, + "learning_rate": 0.0001614931053846634, + "loss": 0.0954, + "step": 37030 + }, + { + "epoch": 0.5777750046796032, + "grad_norm": 1.7900562286376953, + "learning_rate": 0.0001614827062665086, + "loss": 0.1284, + "step": 37040 + }, + { + "epoch": 0.5779309914519248, + "grad_norm": 0.6629091501235962, + "learning_rate": 0.00016147230714835383, + "loss": 0.2657, + "step": 37050 + }, + { + "epoch": 0.5780869782242466, + "grad_norm": 2.502882242202759, + "learning_rate": 0.00016146190803019905, + "loss": 0.207, + "step": 37060 + }, + { + "epoch": 0.5782429649965682, + "grad_norm": 1.486069917678833, + "learning_rate": 0.00016145150891204427, + "loss": 0.1232, + "step": 37070 + }, + { + "epoch": 0.57839895176889, + "grad_norm": 0.13543163239955902, + "learning_rate": 0.0001614411097938895, + "loss": 0.3245, + "step": 37080 + }, + { + "epoch": 0.5785549385412118, + "grad_norm": 3.9703898429870605, + "learning_rate": 0.0001614307106757347, + "loss": 0.3104, + "step": 37090 + }, + { + "epoch": 0.5787109253135334, + "grad_norm": 3.2236313819885254, + "learning_rate": 0.00016142031155757993, + "loss": 0.1739, + "step": 37100 + }, + { + "epoch": 0.5788669120858552, + "grad_norm": 0.3854866325855255, + "learning_rate": 0.00016140991243942515, + "loss": 0.084, + "step": 37110 + }, + { + "epoch": 0.5790228988581768, + "grad_norm": 0.1981169581413269, + "learning_rate": 0.00016139951332127037, + "loss": 0.2367, + "step": 37120 + }, + { + "epoch": 0.5791788856304986, + "grad_norm": 2.7315785884857178, + "learning_rate": 0.0001613891142031156, + "loss": 0.1816, + "step": 37130 + }, + { + "epoch": 0.5793348724028202, + "grad_norm": 0.48251378536224365, + "learning_rate": 0.0001613787150849608, + "loss": 0.2123, + "step": 37140 + }, + { + "epoch": 0.579490859175142, + "grad_norm": 5.0431413650512695, + "learning_rate": 0.00016136831596680603, + "loss": 0.2489, + "step": 37150 + }, + { + "epoch": 0.5796468459474636, + "grad_norm": 0.81981360912323, + "learning_rate": 0.00016135791684865125, + "loss": 0.5235, + "step": 37160 + }, + { + "epoch": 0.5798028327197854, + "grad_norm": 0.5701965689659119, + "learning_rate": 0.00016134751773049647, + "loss": 0.1709, + "step": 37170 + }, + { + "epoch": 0.579958819492107, + "grad_norm": 1.3905616998672485, + "learning_rate": 0.00016133711861234169, + "loss": 0.2641, + "step": 37180 + }, + { + "epoch": 0.5801148062644288, + "grad_norm": 2.281285047531128, + "learning_rate": 0.0001613267194941869, + "loss": 0.3915, + "step": 37190 + }, + { + "epoch": 0.5802707930367504, + "grad_norm": 1.2038309574127197, + "learning_rate": 0.00016131632037603213, + "loss": 0.0881, + "step": 37200 + }, + { + "epoch": 0.5804267798090722, + "grad_norm": 0.3096400201320648, + "learning_rate": 0.00016130592125787735, + "loss": 0.1341, + "step": 37210 + }, + { + "epoch": 0.5805827665813938, + "grad_norm": 0.10219905525445938, + "learning_rate": 0.00016129552213972256, + "loss": 0.0867, + "step": 37220 + }, + { + "epoch": 0.5807387533537156, + "grad_norm": 0.13869646191596985, + "learning_rate": 0.00016128512302156778, + "loss": 0.1621, + "step": 37230 + }, + { + "epoch": 0.5808947401260373, + "grad_norm": 0.7992938160896301, + "learning_rate": 0.000161274723903413, + "loss": 0.275, + "step": 37240 + }, + { + "epoch": 0.581050726898359, + "grad_norm": 0.2635972201824188, + "learning_rate": 0.00016126432478525822, + "loss": 0.2429, + "step": 37250 + }, + { + "epoch": 0.5812067136706808, + "grad_norm": 1.7211238145828247, + "learning_rate": 0.00016125392566710344, + "loss": 0.1541, + "step": 37260 + }, + { + "epoch": 0.5813627004430024, + "grad_norm": 2.2448506355285645, + "learning_rate": 0.00016124352654894866, + "loss": 0.3689, + "step": 37270 + }, + { + "epoch": 0.5815186872153242, + "grad_norm": 1.238708257675171, + "learning_rate": 0.00016123312743079388, + "loss": 0.3292, + "step": 37280 + }, + { + "epoch": 0.5816746739876458, + "grad_norm": 0.5670503377914429, + "learning_rate": 0.0001612227283126391, + "loss": 0.1151, + "step": 37290 + }, + { + "epoch": 0.5818306607599676, + "grad_norm": 2.287304401397705, + "learning_rate": 0.00016121232919448432, + "loss": 0.3558, + "step": 37300 + }, + { + "epoch": 0.5819866475322892, + "grad_norm": 0.8009629845619202, + "learning_rate": 0.00016120193007632954, + "loss": 0.1861, + "step": 37310 + }, + { + "epoch": 0.582142634304611, + "grad_norm": 1.1901466846466064, + "learning_rate": 0.00016119153095817476, + "loss": 0.089, + "step": 37320 + }, + { + "epoch": 0.5822986210769326, + "grad_norm": 1.7198141813278198, + "learning_rate": 0.00016118113184001998, + "loss": 0.2931, + "step": 37330 + }, + { + "epoch": 0.5824546078492544, + "grad_norm": 0.8832749724388123, + "learning_rate": 0.0001611707327218652, + "loss": 0.2342, + "step": 37340 + }, + { + "epoch": 0.582610594621576, + "grad_norm": 1.0623290538787842, + "learning_rate": 0.00016116033360371042, + "loss": 0.2317, + "step": 37350 + }, + { + "epoch": 0.5827665813938978, + "grad_norm": 0.5173699855804443, + "learning_rate": 0.00016114993448555564, + "loss": 0.1299, + "step": 37360 + }, + { + "epoch": 0.5829225681662195, + "grad_norm": 0.05118720978498459, + "learning_rate": 0.00016113953536740086, + "loss": 0.0786, + "step": 37370 + }, + { + "epoch": 0.5830785549385412, + "grad_norm": 2.0060107707977295, + "learning_rate": 0.00016112913624924608, + "loss": 0.266, + "step": 37380 + }, + { + "epoch": 0.5832345417108629, + "grad_norm": 5.121693134307861, + "learning_rate": 0.0001611187371310913, + "loss": 0.1824, + "step": 37390 + }, + { + "epoch": 0.5833905284831846, + "grad_norm": 0.3568089008331299, + "learning_rate": 0.00016110833801293652, + "loss": 0.3189, + "step": 37400 + }, + { + "epoch": 0.5835465152555064, + "grad_norm": 0.8631492853164673, + "learning_rate": 0.00016109793889478174, + "loss": 0.1724, + "step": 37410 + }, + { + "epoch": 0.583702502027828, + "grad_norm": 1.9214404821395874, + "learning_rate": 0.00016108753977662693, + "loss": 0.2094, + "step": 37420 + }, + { + "epoch": 0.5838584888001498, + "grad_norm": 1.9055646657943726, + "learning_rate": 0.00016107714065847218, + "loss": 0.1258, + "step": 37430 + }, + { + "epoch": 0.5840144755724714, + "grad_norm": 0.6614194512367249, + "learning_rate": 0.00016106674154031737, + "loss": 0.1966, + "step": 37440 + }, + { + "epoch": 0.5841704623447932, + "grad_norm": 0.286883682012558, + "learning_rate": 0.00016105634242216262, + "loss": 0.588, + "step": 37450 + }, + { + "epoch": 0.5843264491171148, + "grad_norm": 0.5599405765533447, + "learning_rate": 0.0001610459433040078, + "loss": 0.3414, + "step": 37460 + }, + { + "epoch": 0.5844824358894366, + "grad_norm": 1.5908915996551514, + "learning_rate": 0.00016103554418585306, + "loss": 0.1264, + "step": 37470 + }, + { + "epoch": 0.5846384226617583, + "grad_norm": 1.0729930400848389, + "learning_rate": 0.00016102514506769825, + "loss": 0.4815, + "step": 37480 + }, + { + "epoch": 0.58479440943408, + "grad_norm": 0.8912356495857239, + "learning_rate": 0.0001610147459495435, + "loss": 0.2092, + "step": 37490 + }, + { + "epoch": 0.5849503962064017, + "grad_norm": 0.24116197228431702, + "learning_rate": 0.0001610043468313887, + "loss": 0.1552, + "step": 37500 + }, + { + "epoch": 0.5851063829787234, + "grad_norm": 2.245192527770996, + "learning_rate": 0.00016099394771323393, + "loss": 0.0553, + "step": 37510 + }, + { + "epoch": 0.5852623697510451, + "grad_norm": 0.11538795381784439, + "learning_rate": 0.00016098354859507913, + "loss": 0.0889, + "step": 37520 + }, + { + "epoch": 0.5854183565233668, + "grad_norm": 1.0372689962387085, + "learning_rate": 0.00016097314947692437, + "loss": 0.3909, + "step": 37530 + }, + { + "epoch": 0.5855743432956885, + "grad_norm": 1.1946550607681274, + "learning_rate": 0.00016096275035876957, + "loss": 0.2167, + "step": 37540 + }, + { + "epoch": 0.5857303300680102, + "grad_norm": 1.3474591970443726, + "learning_rate": 0.0001609523512406148, + "loss": 0.1561, + "step": 37550 + }, + { + "epoch": 0.585886316840332, + "grad_norm": 0.008936616592109203, + "learning_rate": 0.00016094195212246, + "loss": 0.3991, + "step": 37560 + }, + { + "epoch": 0.5860423036126536, + "grad_norm": 1.8822320699691772, + "learning_rate": 0.00016093155300430525, + "loss": 0.2438, + "step": 37570 + }, + { + "epoch": 0.5861982903849754, + "grad_norm": 1.1707524061203003, + "learning_rate": 0.00016092115388615044, + "loss": 0.308, + "step": 37580 + }, + { + "epoch": 0.586354277157297, + "grad_norm": 1.679641842842102, + "learning_rate": 0.0001609107547679957, + "loss": 0.159, + "step": 37590 + }, + { + "epoch": 0.5865102639296188, + "grad_norm": 1.6885284185409546, + "learning_rate": 0.00016090035564984088, + "loss": 0.157, + "step": 37600 + }, + { + "epoch": 0.5866662507019405, + "grad_norm": 0.3876996636390686, + "learning_rate": 0.00016088995653168613, + "loss": 0.4067, + "step": 37610 + }, + { + "epoch": 0.5868222374742622, + "grad_norm": 1.387480616569519, + "learning_rate": 0.00016087955741353132, + "loss": 0.1898, + "step": 37620 + }, + { + "epoch": 0.5869782242465839, + "grad_norm": 0.03653861582279205, + "learning_rate": 0.00016086915829537657, + "loss": 0.1652, + "step": 37630 + }, + { + "epoch": 0.5871342110189056, + "grad_norm": 0.6258412003517151, + "learning_rate": 0.00016085875917722176, + "loss": 0.2558, + "step": 37640 + }, + { + "epoch": 0.5872901977912273, + "grad_norm": 0.10519164055585861, + "learning_rate": 0.000160848360059067, + "loss": 0.243, + "step": 37650 + }, + { + "epoch": 0.587446184563549, + "grad_norm": 0.7370131015777588, + "learning_rate": 0.0001608379609409122, + "loss": 0.2343, + "step": 37660 + }, + { + "epoch": 0.5876021713358707, + "grad_norm": 0.6884517073631287, + "learning_rate": 0.00016082756182275745, + "loss": 0.1847, + "step": 37670 + }, + { + "epoch": 0.5877581581081924, + "grad_norm": 0.6326120495796204, + "learning_rate": 0.00016081716270460264, + "loss": 0.2258, + "step": 37680 + }, + { + "epoch": 0.5879141448805141, + "grad_norm": 7.697889804840088, + "learning_rate": 0.0001608067635864479, + "loss": 0.2068, + "step": 37690 + }, + { + "epoch": 0.5880701316528358, + "grad_norm": 1.8753420114517212, + "learning_rate": 0.00016079636446829308, + "loss": 0.3087, + "step": 37700 + }, + { + "epoch": 0.5882261184251576, + "grad_norm": 0.8591625094413757, + "learning_rate": 0.00016078596535013833, + "loss": 0.1082, + "step": 37710 + }, + { + "epoch": 0.5883821051974792, + "grad_norm": 0.6124423146247864, + "learning_rate": 0.00016077556623198352, + "loss": 0.1536, + "step": 37720 + }, + { + "epoch": 0.588538091969801, + "grad_norm": 0.1400165557861328, + "learning_rate": 0.00016076516711382877, + "loss": 0.145, + "step": 37730 + }, + { + "epoch": 0.5886940787421227, + "grad_norm": 1.9879004955291748, + "learning_rate": 0.00016075476799567396, + "loss": 0.4986, + "step": 37740 + }, + { + "epoch": 0.5888500655144444, + "grad_norm": 0.08687327802181244, + "learning_rate": 0.0001607443688775192, + "loss": 0.3365, + "step": 37750 + }, + { + "epoch": 0.5890060522867661, + "grad_norm": 0.3442278206348419, + "learning_rate": 0.0001607339697593644, + "loss": 0.2274, + "step": 37760 + }, + { + "epoch": 0.5891620390590878, + "grad_norm": 0.06124915927648544, + "learning_rate": 0.00016072357064120965, + "loss": 0.1915, + "step": 37770 + }, + { + "epoch": 0.5893180258314095, + "grad_norm": 1.0996816158294678, + "learning_rate": 0.00016071317152305484, + "loss": 0.1573, + "step": 37780 + }, + { + "epoch": 0.5894740126037312, + "grad_norm": 0.14738436043262482, + "learning_rate": 0.00016070277240490008, + "loss": 0.2524, + "step": 37790 + }, + { + "epoch": 0.5896299993760529, + "grad_norm": 2.3905467987060547, + "learning_rate": 0.00016069237328674528, + "loss": 0.1838, + "step": 37800 + }, + { + "epoch": 0.5897859861483746, + "grad_norm": 2.0800178050994873, + "learning_rate": 0.00016068197416859052, + "loss": 0.3749, + "step": 37810 + }, + { + "epoch": 0.5899419729206963, + "grad_norm": 3.254204511642456, + "learning_rate": 0.00016067157505043574, + "loss": 0.3777, + "step": 37820 + }, + { + "epoch": 0.590097959693018, + "grad_norm": 2.8447892665863037, + "learning_rate": 0.00016066117593228096, + "loss": 0.5338, + "step": 37830 + }, + { + "epoch": 0.5902539464653397, + "grad_norm": 0.624733030796051, + "learning_rate": 0.00016065077681412618, + "loss": 0.2331, + "step": 37840 + }, + { + "epoch": 0.5904099332376614, + "grad_norm": 0.6863507628440857, + "learning_rate": 0.0001606403776959714, + "loss": 0.1747, + "step": 37850 + }, + { + "epoch": 0.5905659200099832, + "grad_norm": 0.6785943508148193, + "learning_rate": 0.00016062997857781662, + "loss": 0.2064, + "step": 37860 + }, + { + "epoch": 0.5907219067823049, + "grad_norm": 0.5739938616752625, + "learning_rate": 0.00016061957945966181, + "loss": 0.2368, + "step": 37870 + }, + { + "epoch": 0.5908778935546266, + "grad_norm": 0.0616387277841568, + "learning_rate": 0.00016060918034150706, + "loss": 0.2224, + "step": 37880 + }, + { + "epoch": 0.5910338803269483, + "grad_norm": 2.745957136154175, + "learning_rate": 0.00016059878122335225, + "loss": 0.3161, + "step": 37890 + }, + { + "epoch": 0.59118986709927, + "grad_norm": 2.52472186088562, + "learning_rate": 0.0001605883821051975, + "loss": 0.1713, + "step": 37900 + }, + { + "epoch": 0.5913458538715917, + "grad_norm": 1.5691877603530884, + "learning_rate": 0.0001605779829870427, + "loss": 0.1402, + "step": 37910 + }, + { + "epoch": 0.5915018406439134, + "grad_norm": 1.848261833190918, + "learning_rate": 0.00016056758386888794, + "loss": 0.2305, + "step": 37920 + }, + { + "epoch": 0.5916578274162351, + "grad_norm": 0.36622482538223267, + "learning_rate": 0.00016055718475073313, + "loss": 0.1834, + "step": 37930 + }, + { + "epoch": 0.5918138141885568, + "grad_norm": 0.21201446652412415, + "learning_rate": 0.00016054678563257838, + "loss": 0.4682, + "step": 37940 + }, + { + "epoch": 0.5919698009608785, + "grad_norm": 0.38921502232551575, + "learning_rate": 0.00016053638651442357, + "loss": 0.2892, + "step": 37950 + }, + { + "epoch": 0.5921257877332002, + "grad_norm": 0.20777840912342072, + "learning_rate": 0.00016052598739626882, + "loss": 0.104, + "step": 37960 + }, + { + "epoch": 0.5922817745055219, + "grad_norm": 0.7090150713920593, + "learning_rate": 0.000160515588278114, + "loss": 0.2592, + "step": 37970 + }, + { + "epoch": 0.5924377612778436, + "grad_norm": 1.9939709901809692, + "learning_rate": 0.00016050518915995926, + "loss": 0.247, + "step": 37980 + }, + { + "epoch": 0.5925937480501653, + "grad_norm": 0.8696603178977966, + "learning_rate": 0.00016049479004180445, + "loss": 0.301, + "step": 37990 + }, + { + "epoch": 0.592749734822487, + "grad_norm": 0.4119095206260681, + "learning_rate": 0.0001604843909236497, + "loss": 0.2011, + "step": 38000 + }, + { + "epoch": 0.5929057215948088, + "grad_norm": 0.060975607484579086, + "learning_rate": 0.0001604739918054949, + "loss": 0.2286, + "step": 38010 + }, + { + "epoch": 0.5930617083671305, + "grad_norm": 0.7039875388145447, + "learning_rate": 0.00016046359268734014, + "loss": 0.0999, + "step": 38020 + }, + { + "epoch": 0.5932176951394522, + "grad_norm": 0.28402045369148254, + "learning_rate": 0.00016045319356918533, + "loss": 0.1835, + "step": 38030 + }, + { + "epoch": 0.5933736819117739, + "grad_norm": 2.217988967895508, + "learning_rate": 0.00016044279445103058, + "loss": 0.0785, + "step": 38040 + }, + { + "epoch": 0.5935296686840956, + "grad_norm": 1.4856069087982178, + "learning_rate": 0.00016043239533287577, + "loss": 0.2528, + "step": 38050 + }, + { + "epoch": 0.5936856554564173, + "grad_norm": 0.9875519871711731, + "learning_rate": 0.00016042199621472101, + "loss": 0.4445, + "step": 38060 + }, + { + "epoch": 0.593841642228739, + "grad_norm": 2.773216724395752, + "learning_rate": 0.0001604115970965662, + "loss": 0.5643, + "step": 38070 + }, + { + "epoch": 0.5939976290010607, + "grad_norm": 1.7584549188613892, + "learning_rate": 0.00016040119797841145, + "loss": 0.1446, + "step": 38080 + }, + { + "epoch": 0.5941536157733824, + "grad_norm": 1.2456908226013184, + "learning_rate": 0.00016039079886025665, + "loss": 0.1896, + "step": 38090 + }, + { + "epoch": 0.5943096025457041, + "grad_norm": 0.021098516881465912, + "learning_rate": 0.0001603803997421019, + "loss": 0.0876, + "step": 38100 + }, + { + "epoch": 0.5944655893180258, + "grad_norm": 1.8131749629974365, + "learning_rate": 0.00016037000062394709, + "loss": 0.1639, + "step": 38110 + }, + { + "epoch": 0.5946215760903475, + "grad_norm": 0.41819462180137634, + "learning_rate": 0.00016035960150579233, + "loss": 0.1311, + "step": 38120 + }, + { + "epoch": 0.5947775628626693, + "grad_norm": 0.6691219806671143, + "learning_rate": 0.00016034920238763753, + "loss": 0.1625, + "step": 38130 + }, + { + "epoch": 0.5949335496349909, + "grad_norm": 0.07854852080345154, + "learning_rate": 0.00016033880326948277, + "loss": 0.3177, + "step": 38140 + }, + { + "epoch": 0.5950895364073127, + "grad_norm": 0.7819689512252808, + "learning_rate": 0.00016032840415132796, + "loss": 0.4225, + "step": 38150 + }, + { + "epoch": 0.5952455231796344, + "grad_norm": 5.3962836265563965, + "learning_rate": 0.0001603180050331732, + "loss": 0.123, + "step": 38160 + }, + { + "epoch": 0.5954015099519561, + "grad_norm": 0.7460303902626038, + "learning_rate": 0.0001603076059150184, + "loss": 0.2182, + "step": 38170 + }, + { + "epoch": 0.5955574967242778, + "grad_norm": 0.26433685421943665, + "learning_rate": 0.00016029720679686365, + "loss": 0.1894, + "step": 38180 + }, + { + "epoch": 0.5957134834965995, + "grad_norm": 0.8701320290565491, + "learning_rate": 0.00016028680767870884, + "loss": 0.2137, + "step": 38190 + }, + { + "epoch": 0.5958694702689212, + "grad_norm": 3.9774930477142334, + "learning_rate": 0.0001602764085605541, + "loss": 0.4637, + "step": 38200 + }, + { + "epoch": 0.5960254570412429, + "grad_norm": 0.5801145434379578, + "learning_rate": 0.00016026600944239928, + "loss": 0.3036, + "step": 38210 + }, + { + "epoch": 0.5961814438135646, + "grad_norm": 1.1890723705291748, + "learning_rate": 0.00016025561032424453, + "loss": 0.0988, + "step": 38220 + }, + { + "epoch": 0.5963374305858863, + "grad_norm": 0.18521073460578918, + "learning_rate": 0.00016024521120608972, + "loss": 0.1529, + "step": 38230 + }, + { + "epoch": 0.596493417358208, + "grad_norm": 2.2602827548980713, + "learning_rate": 0.00016023481208793497, + "loss": 0.2815, + "step": 38240 + }, + { + "epoch": 0.5966494041305297, + "grad_norm": 0.4060989022254944, + "learning_rate": 0.00016022441296978016, + "loss": 0.3466, + "step": 38250 + }, + { + "epoch": 0.5968053909028515, + "grad_norm": 1.4743244647979736, + "learning_rate": 0.0001602140138516254, + "loss": 0.3561, + "step": 38260 + }, + { + "epoch": 0.5969613776751731, + "grad_norm": 5.080365180969238, + "learning_rate": 0.0001602036147334706, + "loss": 0.2914, + "step": 38270 + }, + { + "epoch": 0.5971173644474949, + "grad_norm": 0.16149812936782837, + "learning_rate": 0.00016019321561531585, + "loss": 0.1871, + "step": 38280 + }, + { + "epoch": 0.5972733512198165, + "grad_norm": 1.7535364627838135, + "learning_rate": 0.00016018281649716104, + "loss": 0.4905, + "step": 38290 + }, + { + "epoch": 0.5974293379921383, + "grad_norm": 0.4103987216949463, + "learning_rate": 0.00016017241737900629, + "loss": 0.3813, + "step": 38300 + }, + { + "epoch": 0.59758532476446, + "grad_norm": 0.5315369367599487, + "learning_rate": 0.00016016201826085148, + "loss": 0.3897, + "step": 38310 + }, + { + "epoch": 0.5977413115367817, + "grad_norm": 3.4553463459014893, + "learning_rate": 0.0001601516191426967, + "loss": 0.2047, + "step": 38320 + }, + { + "epoch": 0.5978972983091034, + "grad_norm": 2.225468397140503, + "learning_rate": 0.00016014122002454192, + "loss": 0.4028, + "step": 38330 + }, + { + "epoch": 0.5980532850814251, + "grad_norm": 2.000885486602783, + "learning_rate": 0.00016013082090638714, + "loss": 0.391, + "step": 38340 + }, + { + "epoch": 0.5982092718537468, + "grad_norm": 0.7886672616004944, + "learning_rate": 0.00016012042178823236, + "loss": 0.2994, + "step": 38350 + }, + { + "epoch": 0.5983652586260685, + "grad_norm": 0.5279917120933533, + "learning_rate": 0.00016011002267007758, + "loss": 0.1855, + "step": 38360 + }, + { + "epoch": 0.5985212453983902, + "grad_norm": 0.9789942502975464, + "learning_rate": 0.0001600996235519228, + "loss": 0.2122, + "step": 38370 + }, + { + "epoch": 0.5986772321707119, + "grad_norm": 0.6289750933647156, + "learning_rate": 0.00016008922443376802, + "loss": 0.3737, + "step": 38380 + }, + { + "epoch": 0.5988332189430337, + "grad_norm": 1.3074733018875122, + "learning_rate": 0.00016007882531561324, + "loss": 0.2382, + "step": 38390 + }, + { + "epoch": 0.5989892057153553, + "grad_norm": 1.745284914970398, + "learning_rate": 0.00016006842619745846, + "loss": 0.3167, + "step": 38400 + }, + { + "epoch": 0.5991451924876771, + "grad_norm": 0.1694452464580536, + "learning_rate": 0.00016005802707930368, + "loss": 0.1632, + "step": 38410 + }, + { + "epoch": 0.5993011792599987, + "grad_norm": 1.55560302734375, + "learning_rate": 0.0001600476279611489, + "loss": 0.2371, + "step": 38420 + }, + { + "epoch": 0.5994571660323205, + "grad_norm": 1.044997215270996, + "learning_rate": 0.00016003722884299411, + "loss": 0.0896, + "step": 38430 + }, + { + "epoch": 0.5996131528046421, + "grad_norm": 0.25250110030174255, + "learning_rate": 0.00016002682972483933, + "loss": 0.3096, + "step": 38440 + }, + { + "epoch": 0.5997691395769639, + "grad_norm": 0.1631859689950943, + "learning_rate": 0.00016001643060668455, + "loss": 0.2004, + "step": 38450 + }, + { + "epoch": 0.5999251263492856, + "grad_norm": 0.34967318177223206, + "learning_rate": 0.00016000603148852977, + "loss": 0.1476, + "step": 38460 + }, + { + "epoch": 0.6000811131216073, + "grad_norm": 1.8604000806808472, + "learning_rate": 0.000159995632370375, + "loss": 0.1884, + "step": 38470 + }, + { + "epoch": 0.600237099893929, + "grad_norm": 1.626848578453064, + "learning_rate": 0.0001599852332522202, + "loss": 0.5452, + "step": 38480 + }, + { + "epoch": 0.6003930866662507, + "grad_norm": 0.41935989260673523, + "learning_rate": 0.00015997483413406543, + "loss": 0.3047, + "step": 38490 + }, + { + "epoch": 0.6005490734385724, + "grad_norm": 1.355086088180542, + "learning_rate": 0.00015996443501591065, + "loss": 0.1246, + "step": 38500 + }, + { + "epoch": 0.6007050602108941, + "grad_norm": 1.5382829904556274, + "learning_rate": 0.00015995403589775587, + "loss": 0.2324, + "step": 38510 + }, + { + "epoch": 0.6008610469832159, + "grad_norm": 1.0250797271728516, + "learning_rate": 0.0001599436367796011, + "loss": 0.181, + "step": 38520 + }, + { + "epoch": 0.6010170337555375, + "grad_norm": 0.07823364436626434, + "learning_rate": 0.0001599332376614463, + "loss": 0.1492, + "step": 38530 + }, + { + "epoch": 0.6011730205278593, + "grad_norm": 0.07573071867227554, + "learning_rate": 0.00015992283854329153, + "loss": 0.1256, + "step": 38540 + }, + { + "epoch": 0.6013290073001809, + "grad_norm": 0.5608569383621216, + "learning_rate": 0.00015991243942513675, + "loss": 0.378, + "step": 38550 + }, + { + "epoch": 0.6014849940725027, + "grad_norm": 0.7001075744628906, + "learning_rate": 0.00015990204030698197, + "loss": 0.2518, + "step": 38560 + }, + { + "epoch": 0.6016409808448243, + "grad_norm": 4.868145942687988, + "learning_rate": 0.0001598916411888272, + "loss": 0.1857, + "step": 38570 + }, + { + "epoch": 0.6017969676171461, + "grad_norm": 0.05653705075383186, + "learning_rate": 0.0001598812420706724, + "loss": 0.1918, + "step": 38580 + }, + { + "epoch": 0.6019529543894677, + "grad_norm": 0.8556832671165466, + "learning_rate": 0.00015987084295251763, + "loss": 0.1926, + "step": 38590 + }, + { + "epoch": 0.6021089411617895, + "grad_norm": 1.1988335847854614, + "learning_rate": 0.00015986044383436285, + "loss": 0.3884, + "step": 38600 + }, + { + "epoch": 0.6022649279341112, + "grad_norm": 1.5978832244873047, + "learning_rate": 0.00015985004471620807, + "loss": 0.1986, + "step": 38610 + }, + { + "epoch": 0.6024209147064329, + "grad_norm": 1.084174633026123, + "learning_rate": 0.0001598396455980533, + "loss": 0.4602, + "step": 38620 + }, + { + "epoch": 0.6025769014787546, + "grad_norm": 1.5116629600524902, + "learning_rate": 0.0001598292464798985, + "loss": 0.292, + "step": 38630 + }, + { + "epoch": 0.6027328882510763, + "grad_norm": 0.23945322632789612, + "learning_rate": 0.00015981884736174373, + "loss": 0.1904, + "step": 38640 + }, + { + "epoch": 0.602888875023398, + "grad_norm": 0.348156601190567, + "learning_rate": 0.00015980844824358895, + "loss": 0.1442, + "step": 38650 + }, + { + "epoch": 0.6030448617957197, + "grad_norm": 1.876736044883728, + "learning_rate": 0.00015979804912543417, + "loss": 0.248, + "step": 38660 + }, + { + "epoch": 0.6032008485680415, + "grad_norm": 0.2564373016357422, + "learning_rate": 0.00015978765000727939, + "loss": 0.2429, + "step": 38670 + }, + { + "epoch": 0.6033568353403631, + "grad_norm": 1.824837327003479, + "learning_rate": 0.0001597772508891246, + "loss": 0.2593, + "step": 38680 + }, + { + "epoch": 0.6035128221126849, + "grad_norm": 1.6967028379440308, + "learning_rate": 0.00015976685177096985, + "loss": 0.3918, + "step": 38690 + }, + { + "epoch": 0.6036688088850065, + "grad_norm": 1.6629718542099, + "learning_rate": 0.00015975645265281504, + "loss": 0.2529, + "step": 38700 + }, + { + "epoch": 0.6038247956573283, + "grad_norm": 0.19047370553016663, + "learning_rate": 0.0001597460535346603, + "loss": 0.4545, + "step": 38710 + }, + { + "epoch": 0.6039807824296499, + "grad_norm": 0.6910511255264282, + "learning_rate": 0.00015973565441650548, + "loss": 0.2226, + "step": 38720 + }, + { + "epoch": 0.6041367692019717, + "grad_norm": 1.9387328624725342, + "learning_rate": 0.00015972525529835073, + "loss": 0.5561, + "step": 38730 + }, + { + "epoch": 0.6042927559742933, + "grad_norm": 1.4491708278656006, + "learning_rate": 0.00015971485618019592, + "loss": 0.2987, + "step": 38740 + }, + { + "epoch": 0.6044487427466151, + "grad_norm": 0.32534459233283997, + "learning_rate": 0.00015970445706204117, + "loss": 0.1683, + "step": 38750 + }, + { + "epoch": 0.6046047295189368, + "grad_norm": 0.17335011065006256, + "learning_rate": 0.00015969405794388636, + "loss": 0.1554, + "step": 38760 + }, + { + "epoch": 0.6047607162912585, + "grad_norm": 0.39481663703918457, + "learning_rate": 0.00015968365882573158, + "loss": 0.186, + "step": 38770 + }, + { + "epoch": 0.6049167030635803, + "grad_norm": 0.10422719269990921, + "learning_rate": 0.0001596732597075768, + "loss": 0.1599, + "step": 38780 + }, + { + "epoch": 0.6050726898359019, + "grad_norm": 2.288064479827881, + "learning_rate": 0.00015966286058942202, + "loss": 0.1944, + "step": 38790 + }, + { + "epoch": 0.6052286766082237, + "grad_norm": 1.0062401294708252, + "learning_rate": 0.00015965246147126724, + "loss": 0.176, + "step": 38800 + }, + { + "epoch": 0.6053846633805453, + "grad_norm": 0.21470843255519867, + "learning_rate": 0.00015964206235311246, + "loss": 0.183, + "step": 38810 + }, + { + "epoch": 0.6055406501528671, + "grad_norm": 2.1730270385742188, + "learning_rate": 0.00015963166323495768, + "loss": 0.1788, + "step": 38820 + }, + { + "epoch": 0.6056966369251887, + "grad_norm": 0.6188749074935913, + "learning_rate": 0.0001596212641168029, + "loss": 0.2746, + "step": 38830 + }, + { + "epoch": 0.6058526236975105, + "grad_norm": 1.283058524131775, + "learning_rate": 0.00015961086499864812, + "loss": 0.1066, + "step": 38840 + }, + { + "epoch": 0.6060086104698321, + "grad_norm": 1.1469789743423462, + "learning_rate": 0.00015960046588049334, + "loss": 0.0692, + "step": 38850 + }, + { + "epoch": 0.6061645972421539, + "grad_norm": 1.9450418949127197, + "learning_rate": 0.00015959006676233856, + "loss": 0.2154, + "step": 38860 + }, + { + "epoch": 0.6063205840144755, + "grad_norm": 7.193401336669922, + "learning_rate": 0.00015957966764418378, + "loss": 0.218, + "step": 38870 + }, + { + "epoch": 0.6064765707867973, + "grad_norm": 3.5315003395080566, + "learning_rate": 0.000159569268526029, + "loss": 0.312, + "step": 38880 + }, + { + "epoch": 0.6066325575591189, + "grad_norm": 1.2873650789260864, + "learning_rate": 0.00015955886940787422, + "loss": 0.2587, + "step": 38890 + }, + { + "epoch": 0.6067885443314407, + "grad_norm": 1.1193941831588745, + "learning_rate": 0.00015954847028971944, + "loss": 0.2358, + "step": 38900 + }, + { + "epoch": 0.6069445311037625, + "grad_norm": 1.0471165180206299, + "learning_rate": 0.00015953807117156466, + "loss": 0.2639, + "step": 38910 + }, + { + "epoch": 0.6071005178760841, + "grad_norm": 1.7760034799575806, + "learning_rate": 0.00015952767205340988, + "loss": 0.248, + "step": 38920 + }, + { + "epoch": 0.6072565046484059, + "grad_norm": 0.050742994993925095, + "learning_rate": 0.0001595172729352551, + "loss": 0.1287, + "step": 38930 + }, + { + "epoch": 0.6074124914207275, + "grad_norm": 0.03016967698931694, + "learning_rate": 0.00015950687381710032, + "loss": 0.1843, + "step": 38940 + }, + { + "epoch": 0.6075684781930493, + "grad_norm": 0.6029847860336304, + "learning_rate": 0.00015949647469894554, + "loss": 0.2294, + "step": 38950 + }, + { + "epoch": 0.6077244649653709, + "grad_norm": 0.2955610454082489, + "learning_rate": 0.00015948607558079076, + "loss": 0.3336, + "step": 38960 + }, + { + "epoch": 0.6078804517376927, + "grad_norm": 1.9104161262512207, + "learning_rate": 0.00015947567646263597, + "loss": 0.189, + "step": 38970 + }, + { + "epoch": 0.6080364385100143, + "grad_norm": 0.09197711199522018, + "learning_rate": 0.0001594652773444812, + "loss": 0.1612, + "step": 38980 + }, + { + "epoch": 0.6081924252823361, + "grad_norm": 1.4233628511428833, + "learning_rate": 0.00015945487822632641, + "loss": 0.1521, + "step": 38990 + }, + { + "epoch": 0.6083484120546577, + "grad_norm": 9.787493705749512, + "learning_rate": 0.00015944447910817163, + "loss": 0.135, + "step": 39000 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.374935773184e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..21bff82a2f1f4632379e39f0f932c70dbd930656 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10dcaa368a8937c092fc960406694f8e2259e21fffd4143fbfd44a5938838906 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..4be104714270c91f9d31a8290c0e333616be37c1 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aece7368f6904a776d524b39842fdee1785287eaf16fd771777d5c2111745b8f +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..faeb8f70e1a9a611ae50902af55ffc4dfc8170ca Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..a85b4e15f770e39a74184c0144ff439c74da670a Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..218b9a5ea2bfd779ba84ad744ab52bb550eb4adb --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/trainer_state.json @@ -0,0 +1,27684 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.6161477506707431, + "eval_steps": 500, + "global_step": 39500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + }, + { + "epoch": 0.5071129968178698, + "grad_norm": 1.0075199604034424, + "learning_rate": 0.00016619350679062414, + "loss": 0.2569, + "step": 32510 + }, + { + "epoch": 0.5072689835901916, + "grad_norm": 1.3489066362380981, + "learning_rate": 0.0001661831076724694, + "loss": 0.2565, + "step": 32520 + }, + { + "epoch": 0.5074249703625132, + "grad_norm": 0.2534058094024658, + "learning_rate": 0.00016617270855431458, + "loss": 0.2292, + "step": 32530 + }, + { + "epoch": 0.507580957134835, + "grad_norm": 0.31723281741142273, + "learning_rate": 0.00016616230943615983, + "loss": 0.1467, + "step": 32540 + }, + { + "epoch": 0.5077369439071566, + "grad_norm": 0.29659849405288696, + "learning_rate": 0.00016615191031800502, + "loss": 0.2768, + "step": 32550 + }, + { + "epoch": 0.5078929306794784, + "grad_norm": 1.5246435403823853, + "learning_rate": 0.00016614151119985027, + "loss": 0.216, + "step": 32560 + }, + { + "epoch": 0.5080489174518001, + "grad_norm": 0.6026854515075684, + "learning_rate": 0.00016613111208169546, + "loss": 0.2277, + "step": 32570 + }, + { + "epoch": 0.5082049042241218, + "grad_norm": 0.41994890570640564, + "learning_rate": 0.0001661207129635407, + "loss": 0.2433, + "step": 32580 + }, + { + "epoch": 0.5083608909964435, + "grad_norm": 2.084599733352661, + "learning_rate": 0.00016611031384538593, + "loss": 0.4081, + "step": 32590 + }, + { + "epoch": 0.5085168777687652, + "grad_norm": 0.2945229709148407, + "learning_rate": 0.00016609991472723115, + "loss": 0.1329, + "step": 32600 + }, + { + "epoch": 0.508672864541087, + "grad_norm": 1.4981423616409302, + "learning_rate": 0.00016608951560907636, + "loss": 0.2855, + "step": 32610 + }, + { + "epoch": 0.5088288513134086, + "grad_norm": 0.961737871170044, + "learning_rate": 0.00016607911649092158, + "loss": 0.1654, + "step": 32620 + }, + { + "epoch": 0.5089848380857304, + "grad_norm": 1.49734628200531, + "learning_rate": 0.0001660687173727668, + "loss": 0.2081, + "step": 32630 + }, + { + "epoch": 0.509140824858052, + "grad_norm": 2.538135051727295, + "learning_rate": 0.00016605831825461202, + "loss": 0.3023, + "step": 32640 + }, + { + "epoch": 0.5092968116303738, + "grad_norm": 0.2981272041797638, + "learning_rate": 0.00016604791913645724, + "loss": 0.214, + "step": 32650 + }, + { + "epoch": 0.5094527984026954, + "grad_norm": 1.8863650560379028, + "learning_rate": 0.00016603752001830246, + "loss": 0.3363, + "step": 32660 + }, + { + "epoch": 0.5096087851750172, + "grad_norm": 0.022720983251929283, + "learning_rate": 0.00016602712090014768, + "loss": 0.2463, + "step": 32670 + }, + { + "epoch": 0.5097647719473388, + "grad_norm": 1.154977798461914, + "learning_rate": 0.0001660167217819929, + "loss": 0.4484, + "step": 32680 + }, + { + "epoch": 0.5099207587196606, + "grad_norm": 0.9470499157905579, + "learning_rate": 0.00016600632266383812, + "loss": 0.1196, + "step": 32690 + }, + { + "epoch": 0.5100767454919822, + "grad_norm": 0.19955401122570038, + "learning_rate": 0.00016599592354568334, + "loss": 0.1762, + "step": 32700 + }, + { + "epoch": 0.510232732264304, + "grad_norm": 0.4098282754421234, + "learning_rate": 0.00016598552442752856, + "loss": 0.2739, + "step": 32710 + }, + { + "epoch": 0.5103887190366257, + "grad_norm": 0.21697258949279785, + "learning_rate": 0.00016597512530937378, + "loss": 0.2555, + "step": 32720 + }, + { + "epoch": 0.5105447058089474, + "grad_norm": 4.894479751586914, + "learning_rate": 0.000165964726191219, + "loss": 0.3137, + "step": 32730 + }, + { + "epoch": 0.5107006925812692, + "grad_norm": 0.8396711349487305, + "learning_rate": 0.00016595432707306422, + "loss": 0.3105, + "step": 32740 + }, + { + "epoch": 0.5108566793535908, + "grad_norm": 1.5494674444198608, + "learning_rate": 0.00016594392795490944, + "loss": 0.2271, + "step": 32750 + }, + { + "epoch": 0.5110126661259126, + "grad_norm": 1.3409888744354248, + "learning_rate": 0.00016593352883675466, + "loss": 0.1796, + "step": 32760 + }, + { + "epoch": 0.5111686528982342, + "grad_norm": 1.7590911388397217, + "learning_rate": 0.00016592312971859988, + "loss": 0.2185, + "step": 32770 + }, + { + "epoch": 0.511324639670556, + "grad_norm": 0.7350002527236938, + "learning_rate": 0.0001659127306004451, + "loss": 0.1911, + "step": 32780 + }, + { + "epoch": 0.5114806264428776, + "grad_norm": 0.7176411747932434, + "learning_rate": 0.00016590233148229032, + "loss": 0.6918, + "step": 32790 + }, + { + "epoch": 0.5116366132151994, + "grad_norm": 2.5005321502685547, + "learning_rate": 0.00016589193236413554, + "loss": 0.2329, + "step": 32800 + }, + { + "epoch": 0.511792599987521, + "grad_norm": 1.3916250467300415, + "learning_rate": 0.00016588153324598076, + "loss": 0.2102, + "step": 32810 + }, + { + "epoch": 0.5119485867598428, + "grad_norm": 0.721671998500824, + "learning_rate": 0.00016587113412782598, + "loss": 0.2965, + "step": 32820 + }, + { + "epoch": 0.5121045735321644, + "grad_norm": 0.6120114326477051, + "learning_rate": 0.0001658607350096712, + "loss": 0.2661, + "step": 32830 + }, + { + "epoch": 0.5122605603044862, + "grad_norm": 0.29123347997665405, + "learning_rate": 0.00016585033589151642, + "loss": 0.2455, + "step": 32840 + }, + { + "epoch": 0.5124165470768078, + "grad_norm": 2.7300868034362793, + "learning_rate": 0.00016583993677336164, + "loss": 0.2049, + "step": 32850 + }, + { + "epoch": 0.5125725338491296, + "grad_norm": 1.801928997039795, + "learning_rate": 0.00016582953765520686, + "loss": 0.1725, + "step": 32860 + }, + { + "epoch": 0.5127285206214514, + "grad_norm": 0.6406471729278564, + "learning_rate": 0.00016581913853705208, + "loss": 0.164, + "step": 32870 + }, + { + "epoch": 0.512884507393773, + "grad_norm": 0.9627325534820557, + "learning_rate": 0.0001658087394188973, + "loss": 0.1728, + "step": 32880 + }, + { + "epoch": 0.5130404941660948, + "grad_norm": 1.5446962118148804, + "learning_rate": 0.00016579834030074251, + "loss": 0.2345, + "step": 32890 + }, + { + "epoch": 0.5131964809384164, + "grad_norm": 0.28092706203460693, + "learning_rate": 0.0001657879411825877, + "loss": 0.2124, + "step": 32900 + }, + { + "epoch": 0.5133524677107382, + "grad_norm": 1.7751421928405762, + "learning_rate": 0.00016577754206443295, + "loss": 0.2367, + "step": 32910 + }, + { + "epoch": 0.5135084544830598, + "grad_norm": 1.6964634656906128, + "learning_rate": 0.00016576714294627815, + "loss": 0.2361, + "step": 32920 + }, + { + "epoch": 0.5136644412553816, + "grad_norm": 0.06759823113679886, + "learning_rate": 0.0001657567438281234, + "loss": 0.2329, + "step": 32930 + }, + { + "epoch": 0.5138204280277032, + "grad_norm": 2.2519280910491943, + "learning_rate": 0.00016574634470996859, + "loss": 0.1079, + "step": 32940 + }, + { + "epoch": 0.513976414800025, + "grad_norm": 3.167576551437378, + "learning_rate": 0.00016573594559181383, + "loss": 0.1764, + "step": 32950 + }, + { + "epoch": 0.5141324015723466, + "grad_norm": 1.6734397411346436, + "learning_rate": 0.00016572554647365903, + "loss": 0.2878, + "step": 32960 + }, + { + "epoch": 0.5142883883446684, + "grad_norm": 0.35627374053001404, + "learning_rate": 0.00016571514735550427, + "loss": 0.2337, + "step": 32970 + }, + { + "epoch": 0.51444437511699, + "grad_norm": 1.6321039199829102, + "learning_rate": 0.00016570474823734946, + "loss": 0.2551, + "step": 32980 + }, + { + "epoch": 0.5146003618893118, + "grad_norm": 1.352982997894287, + "learning_rate": 0.0001656943491191947, + "loss": 0.4467, + "step": 32990 + }, + { + "epoch": 0.5147563486616334, + "grad_norm": 0.3558129072189331, + "learning_rate": 0.0001656839500010399, + "loss": 0.1913, + "step": 33000 + }, + { + "epoch": 0.5149123354339552, + "grad_norm": 0.041972286999225616, + "learning_rate": 0.00016567355088288515, + "loss": 0.1704, + "step": 33010 + }, + { + "epoch": 0.515068322206277, + "grad_norm": 1.4168939590454102, + "learning_rate": 0.00016566315176473034, + "loss": 0.4866, + "step": 33020 + }, + { + "epoch": 0.5152243089785986, + "grad_norm": 3.0042929649353027, + "learning_rate": 0.0001656527526465756, + "loss": 0.2473, + "step": 33030 + }, + { + "epoch": 0.5153802957509204, + "grad_norm": 1.027274489402771, + "learning_rate": 0.00016564235352842078, + "loss": 0.436, + "step": 33040 + }, + { + "epoch": 0.515536282523242, + "grad_norm": 1.2330195903778076, + "learning_rate": 0.00016563195441026603, + "loss": 0.4615, + "step": 33050 + }, + { + "epoch": 0.5156922692955638, + "grad_norm": 2.7145066261291504, + "learning_rate": 0.00016562155529211122, + "loss": 0.278, + "step": 33060 + }, + { + "epoch": 0.5158482560678854, + "grad_norm": 1.1273893117904663, + "learning_rate": 0.00016561115617395647, + "loss": 0.3767, + "step": 33070 + }, + { + "epoch": 0.5160042428402072, + "grad_norm": 2.466265916824341, + "learning_rate": 0.00016560075705580166, + "loss": 0.1879, + "step": 33080 + }, + { + "epoch": 0.5161602296125288, + "grad_norm": 0.3759515583515167, + "learning_rate": 0.0001655903579376469, + "loss": 0.4004, + "step": 33090 + }, + { + "epoch": 0.5163162163848506, + "grad_norm": 1.8821336030960083, + "learning_rate": 0.0001655799588194921, + "loss": 0.1622, + "step": 33100 + }, + { + "epoch": 0.5164722031571722, + "grad_norm": 0.8162096738815308, + "learning_rate": 0.00016556955970133735, + "loss": 0.1341, + "step": 33110 + }, + { + "epoch": 0.516628189929494, + "grad_norm": 0.008498801849782467, + "learning_rate": 0.00016555916058318254, + "loss": 0.2546, + "step": 33120 + }, + { + "epoch": 0.5167841767018156, + "grad_norm": 0.17536675930023193, + "learning_rate": 0.00016554876146502779, + "loss": 0.2934, + "step": 33130 + }, + { + "epoch": 0.5169401634741374, + "grad_norm": 3.539898633956909, + "learning_rate": 0.00016553836234687298, + "loss": 0.4959, + "step": 33140 + }, + { + "epoch": 0.517096150246459, + "grad_norm": 2.190887212753296, + "learning_rate": 0.00016552796322871823, + "loss": 0.2914, + "step": 33150 + }, + { + "epoch": 0.5172521370187808, + "grad_norm": 1.6388943195343018, + "learning_rate": 0.00016551756411056342, + "loss": 0.1317, + "step": 33160 + }, + { + "epoch": 0.5174081237911026, + "grad_norm": 0.33063387870788574, + "learning_rate": 0.00016550716499240866, + "loss": 0.4258, + "step": 33170 + }, + { + "epoch": 0.5175641105634242, + "grad_norm": 0.5168288946151733, + "learning_rate": 0.00016549676587425386, + "loss": 0.2548, + "step": 33180 + }, + { + "epoch": 0.517720097335746, + "grad_norm": 3.9570281505584717, + "learning_rate": 0.0001654863667560991, + "loss": 0.3179, + "step": 33190 + }, + { + "epoch": 0.5178760841080676, + "grad_norm": 1.127255916595459, + "learning_rate": 0.0001654759676379443, + "loss": 0.233, + "step": 33200 + }, + { + "epoch": 0.5180320708803894, + "grad_norm": 0.36299410462379456, + "learning_rate": 0.00016546556851978954, + "loss": 0.4614, + "step": 33210 + }, + { + "epoch": 0.518188057652711, + "grad_norm": 2.914668321609497, + "learning_rate": 0.00016545516940163474, + "loss": 0.1603, + "step": 33220 + }, + { + "epoch": 0.5183440444250328, + "grad_norm": 6.991094589233398, + "learning_rate": 0.00016544477028347998, + "loss": 0.2259, + "step": 33230 + }, + { + "epoch": 0.5185000311973544, + "grad_norm": 2.0594241619110107, + "learning_rate": 0.00016543437116532517, + "loss": 0.1885, + "step": 33240 + }, + { + "epoch": 0.5186560179696762, + "grad_norm": 0.5594693422317505, + "learning_rate": 0.00016542397204717042, + "loss": 0.4333, + "step": 33250 + }, + { + "epoch": 0.5188120047419978, + "grad_norm": 4.314854621887207, + "learning_rate": 0.00016541357292901561, + "loss": 0.2103, + "step": 33260 + }, + { + "epoch": 0.5189679915143196, + "grad_norm": 0.5609000325202942, + "learning_rate": 0.00016540317381086086, + "loss": 0.2498, + "step": 33270 + }, + { + "epoch": 0.5191239782866413, + "grad_norm": 3.680088758468628, + "learning_rate": 0.00016539277469270605, + "loss": 0.3465, + "step": 33280 + }, + { + "epoch": 0.519279965058963, + "grad_norm": 1.4993672370910645, + "learning_rate": 0.0001653823755745513, + "loss": 0.3272, + "step": 33290 + }, + { + "epoch": 0.5194359518312847, + "grad_norm": 0.18005084991455078, + "learning_rate": 0.0001653719764563965, + "loss": 0.1706, + "step": 33300 + }, + { + "epoch": 0.5195919386036064, + "grad_norm": 0.5105524659156799, + "learning_rate": 0.00016536157733824174, + "loss": 0.0937, + "step": 33310 + }, + { + "epoch": 0.5197479253759281, + "grad_norm": 1.2806589603424072, + "learning_rate": 0.00016535117822008693, + "loss": 0.1578, + "step": 33320 + }, + { + "epoch": 0.5199039121482498, + "grad_norm": 0.22015990316867828, + "learning_rate": 0.00016534077910193218, + "loss": 0.2081, + "step": 33330 + }, + { + "epoch": 0.5200598989205716, + "grad_norm": 0.34638991951942444, + "learning_rate": 0.00016533037998377737, + "loss": 0.2389, + "step": 33340 + }, + { + "epoch": 0.5202158856928932, + "grad_norm": 0.8709302544593811, + "learning_rate": 0.0001653199808656226, + "loss": 0.2112, + "step": 33350 + }, + { + "epoch": 0.520371872465215, + "grad_norm": 1.2309521436691284, + "learning_rate": 0.0001653095817474678, + "loss": 0.3448, + "step": 33360 + }, + { + "epoch": 0.5205278592375366, + "grad_norm": 1.3155115842819214, + "learning_rate": 0.00016529918262931303, + "loss": 0.1572, + "step": 33370 + }, + { + "epoch": 0.5206838460098584, + "grad_norm": 0.358751118183136, + "learning_rate": 0.00016528878351115825, + "loss": 0.3011, + "step": 33380 + }, + { + "epoch": 0.52083983278218, + "grad_norm": 3.800628662109375, + "learning_rate": 0.00016527838439300347, + "loss": 0.458, + "step": 33390 + }, + { + "epoch": 0.5209958195545018, + "grad_norm": 1.1253124475479126, + "learning_rate": 0.0001652679852748487, + "loss": 0.1238, + "step": 33400 + }, + { + "epoch": 0.5211518063268235, + "grad_norm": 0.46979108452796936, + "learning_rate": 0.0001652575861566939, + "loss": 0.0745, + "step": 33410 + }, + { + "epoch": 0.5213077930991452, + "grad_norm": 7.921492576599121, + "learning_rate": 0.00016524718703853913, + "loss": 0.4004, + "step": 33420 + }, + { + "epoch": 0.5214637798714669, + "grad_norm": 1.643127679824829, + "learning_rate": 0.00016523678792038435, + "loss": 0.2858, + "step": 33430 + }, + { + "epoch": 0.5216197666437886, + "grad_norm": 7.096631050109863, + "learning_rate": 0.0001652263888022296, + "loss": 0.4735, + "step": 33440 + }, + { + "epoch": 0.5217757534161103, + "grad_norm": 0.7280219197273254, + "learning_rate": 0.0001652159896840748, + "loss": 0.1841, + "step": 33450 + }, + { + "epoch": 0.521931740188432, + "grad_norm": 0.9415522813796997, + "learning_rate": 0.00016520559056592003, + "loss": 0.2169, + "step": 33460 + }, + { + "epoch": 0.5220877269607537, + "grad_norm": 2.264862060546875, + "learning_rate": 0.00016519519144776523, + "loss": 0.1902, + "step": 33470 + }, + { + "epoch": 0.5222437137330754, + "grad_norm": 2.4566822052001953, + "learning_rate": 0.00016518479232961047, + "loss": 0.3329, + "step": 33480 + }, + { + "epoch": 0.5223997005053972, + "grad_norm": 0.7330378890037537, + "learning_rate": 0.00016517439321145567, + "loss": 0.2341, + "step": 33490 + }, + { + "epoch": 0.5225556872777188, + "grad_norm": 1.6409542560577393, + "learning_rate": 0.0001651639940933009, + "loss": 0.3077, + "step": 33500 + }, + { + "epoch": 0.5227116740500406, + "grad_norm": 1.0647432804107666, + "learning_rate": 0.0001651535949751461, + "loss": 0.1935, + "step": 33510 + }, + { + "epoch": 0.5228676608223622, + "grad_norm": 0.507420539855957, + "learning_rate": 0.00016514319585699135, + "loss": 0.1314, + "step": 33520 + }, + { + "epoch": 0.523023647594684, + "grad_norm": 1.0064164400100708, + "learning_rate": 0.00016513279673883654, + "loss": 0.344, + "step": 33530 + }, + { + "epoch": 0.5231796343670057, + "grad_norm": 1.1936209201812744, + "learning_rate": 0.0001651223976206818, + "loss": 0.1948, + "step": 33540 + }, + { + "epoch": 0.5233356211393274, + "grad_norm": 0.04992926865816116, + "learning_rate": 0.00016511199850252698, + "loss": 0.1363, + "step": 33550 + }, + { + "epoch": 0.5234916079116491, + "grad_norm": 1.8256275653839111, + "learning_rate": 0.00016510159938437223, + "loss": 0.443, + "step": 33560 + }, + { + "epoch": 0.5236475946839708, + "grad_norm": 4.914908409118652, + "learning_rate": 0.00016509120026621742, + "loss": 0.4654, + "step": 33570 + }, + { + "epoch": 0.5238035814562925, + "grad_norm": 1.1625486612319946, + "learning_rate": 0.00016508080114806267, + "loss": 0.1974, + "step": 33580 + }, + { + "epoch": 0.5239595682286142, + "grad_norm": 6.6192522048950195, + "learning_rate": 0.00016507040202990786, + "loss": 0.2131, + "step": 33590 + }, + { + "epoch": 0.5241155550009359, + "grad_norm": 0.7369065880775452, + "learning_rate": 0.0001650600029117531, + "loss": 0.2371, + "step": 33600 + }, + { + "epoch": 0.5242715417732576, + "grad_norm": 1.5238152742385864, + "learning_rate": 0.0001650496037935983, + "loss": 0.1557, + "step": 33610 + }, + { + "epoch": 0.5244275285455793, + "grad_norm": 1.0418007373809814, + "learning_rate": 0.00016503920467544355, + "loss": 0.1878, + "step": 33620 + }, + { + "epoch": 0.524583515317901, + "grad_norm": 0.790117084980011, + "learning_rate": 0.00016502880555728874, + "loss": 0.2195, + "step": 33630 + }, + { + "epoch": 0.5247395020902228, + "grad_norm": 1.6712257862091064, + "learning_rate": 0.000165018406439134, + "loss": 0.1602, + "step": 33640 + }, + { + "epoch": 0.5248954888625444, + "grad_norm": 0.19236230850219727, + "learning_rate": 0.00016500800732097918, + "loss": 0.2526, + "step": 33650 + }, + { + "epoch": 0.5250514756348662, + "grad_norm": 1.3519701957702637, + "learning_rate": 0.00016499760820282443, + "loss": 0.4686, + "step": 33660 + }, + { + "epoch": 0.5252074624071879, + "grad_norm": 1.694342851638794, + "learning_rate": 0.00016498720908466962, + "loss": 0.1859, + "step": 33670 + }, + { + "epoch": 0.5253634491795096, + "grad_norm": 5.225239276885986, + "learning_rate": 0.00016497680996651487, + "loss": 0.2944, + "step": 33680 + }, + { + "epoch": 0.5255194359518313, + "grad_norm": 2.0208842754364014, + "learning_rate": 0.00016496641084836006, + "loss": 0.2421, + "step": 33690 + }, + { + "epoch": 0.525675422724153, + "grad_norm": 0.7954996228218079, + "learning_rate": 0.0001649560117302053, + "loss": 0.2853, + "step": 33700 + }, + { + "epoch": 0.5258314094964747, + "grad_norm": 2.296086072921753, + "learning_rate": 0.0001649456126120505, + "loss": 0.1333, + "step": 33710 + }, + { + "epoch": 0.5259873962687964, + "grad_norm": 1.1779128313064575, + "learning_rate": 0.00016493521349389574, + "loss": 0.1066, + "step": 33720 + }, + { + "epoch": 0.5261433830411181, + "grad_norm": 0.1756065934896469, + "learning_rate": 0.00016492481437574094, + "loss": 0.1352, + "step": 33730 + }, + { + "epoch": 0.5262993698134398, + "grad_norm": 0.13100725412368774, + "learning_rate": 0.00016491441525758618, + "loss": 0.2399, + "step": 33740 + }, + { + "epoch": 0.5264553565857615, + "grad_norm": 5.532008171081543, + "learning_rate": 0.00016490401613943138, + "loss": 0.2896, + "step": 33750 + }, + { + "epoch": 0.5266113433580832, + "grad_norm": 1.319886565208435, + "learning_rate": 0.00016489361702127662, + "loss": 0.3275, + "step": 33760 + }, + { + "epoch": 0.5267673301304049, + "grad_norm": 1.5550974607467651, + "learning_rate": 0.00016488321790312182, + "loss": 0.2677, + "step": 33770 + }, + { + "epoch": 0.5269233169027266, + "grad_norm": 1.8936737775802612, + "learning_rate": 0.00016487281878496706, + "loss": 0.1955, + "step": 33780 + }, + { + "epoch": 0.5270793036750484, + "grad_norm": 0.3653401732444763, + "learning_rate": 0.00016486241966681226, + "loss": 0.0723, + "step": 33790 + }, + { + "epoch": 0.52723529044737, + "grad_norm": 2.861341714859009, + "learning_rate": 0.00016485202054865747, + "loss": 0.2412, + "step": 33800 + }, + { + "epoch": 0.5273912772196918, + "grad_norm": 1.5291428565979004, + "learning_rate": 0.0001648416214305027, + "loss": 0.0871, + "step": 33810 + }, + { + "epoch": 0.5275472639920135, + "grad_norm": 1.0372581481933594, + "learning_rate": 0.00016483122231234791, + "loss": 0.4705, + "step": 33820 + }, + { + "epoch": 0.5277032507643352, + "grad_norm": 1.1943141222000122, + "learning_rate": 0.00016482082319419313, + "loss": 0.2848, + "step": 33830 + }, + { + "epoch": 0.5278592375366569, + "grad_norm": 1.9008225202560425, + "learning_rate": 0.00016481042407603835, + "loss": 0.139, + "step": 33840 + }, + { + "epoch": 0.5280152243089786, + "grad_norm": 2.132089138031006, + "learning_rate": 0.00016480002495788357, + "loss": 0.2119, + "step": 33850 + }, + { + "epoch": 0.5281712110813003, + "grad_norm": 0.24524426460266113, + "learning_rate": 0.0001647896258397288, + "loss": 0.1643, + "step": 33860 + }, + { + "epoch": 0.528327197853622, + "grad_norm": 1.6469637155532837, + "learning_rate": 0.000164779226721574, + "loss": 0.363, + "step": 33870 + }, + { + "epoch": 0.5284831846259437, + "grad_norm": 0.8767328858375549, + "learning_rate": 0.00016476882760341923, + "loss": 0.2632, + "step": 33880 + }, + { + "epoch": 0.5286391713982654, + "grad_norm": 0.06347586214542389, + "learning_rate": 0.00016475842848526445, + "loss": 0.3204, + "step": 33890 + }, + { + "epoch": 0.5287951581705871, + "grad_norm": 0.09782540798187256, + "learning_rate": 0.00016474802936710967, + "loss": 0.2211, + "step": 33900 + }, + { + "epoch": 0.5289511449429088, + "grad_norm": 3.2998859882354736, + "learning_rate": 0.0001647376302489549, + "loss": 0.265, + "step": 33910 + }, + { + "epoch": 0.5291071317152305, + "grad_norm": 0.43594226241111755, + "learning_rate": 0.0001647272311308001, + "loss": 0.194, + "step": 33920 + }, + { + "epoch": 0.5292631184875523, + "grad_norm": 1.5166605710983276, + "learning_rate": 0.00016471683201264533, + "loss": 0.2675, + "step": 33930 + }, + { + "epoch": 0.529419105259874, + "grad_norm": 0.6056640148162842, + "learning_rate": 0.00016470643289449055, + "loss": 0.0576, + "step": 33940 + }, + { + "epoch": 0.5295750920321957, + "grad_norm": 0.25410348176956177, + "learning_rate": 0.00016469603377633577, + "loss": 0.1124, + "step": 33950 + }, + { + "epoch": 0.5297310788045174, + "grad_norm": 1.770642876625061, + "learning_rate": 0.000164685634658181, + "loss": 0.3295, + "step": 33960 + }, + { + "epoch": 0.5298870655768391, + "grad_norm": 0.0607205331325531, + "learning_rate": 0.0001646752355400262, + "loss": 0.2369, + "step": 33970 + }, + { + "epoch": 0.5300430523491608, + "grad_norm": 0.5557095408439636, + "learning_rate": 0.00016466483642187143, + "loss": 0.1681, + "step": 33980 + }, + { + "epoch": 0.5301990391214825, + "grad_norm": 0.5192957520484924, + "learning_rate": 0.00016465443730371665, + "loss": 0.2649, + "step": 33990 + }, + { + "epoch": 0.5303550258938042, + "grad_norm": 0.04804835096001625, + "learning_rate": 0.00016464403818556187, + "loss": 0.2639, + "step": 34000 + }, + { + "epoch": 0.5305110126661259, + "grad_norm": 0.02673129364848137, + "learning_rate": 0.0001646336390674071, + "loss": 0.169, + "step": 34010 + }, + { + "epoch": 0.5306669994384476, + "grad_norm": 1.0084244012832642, + "learning_rate": 0.0001646232399492523, + "loss": 0.353, + "step": 34020 + }, + { + "epoch": 0.5308229862107693, + "grad_norm": 2.2202091217041016, + "learning_rate": 0.00016461284083109753, + "loss": 0.3206, + "step": 34030 + }, + { + "epoch": 0.530978972983091, + "grad_norm": 0.5573744773864746, + "learning_rate": 0.00016460244171294275, + "loss": 0.231, + "step": 34040 + }, + { + "epoch": 0.5311349597554127, + "grad_norm": 0.39700084924697876, + "learning_rate": 0.00016459204259478797, + "loss": 0.1412, + "step": 34050 + }, + { + "epoch": 0.5312909465277345, + "grad_norm": 2.582963228225708, + "learning_rate": 0.00016458164347663319, + "loss": 0.3334, + "step": 34060 + }, + { + "epoch": 0.5314469333000561, + "grad_norm": 0.22781169414520264, + "learning_rate": 0.0001645712443584784, + "loss": 0.2767, + "step": 34070 + }, + { + "epoch": 0.5316029200723779, + "grad_norm": 1.2051042318344116, + "learning_rate": 0.00016456084524032362, + "loss": 0.2045, + "step": 34080 + }, + { + "epoch": 0.5317589068446996, + "grad_norm": 0.42760100960731506, + "learning_rate": 0.00016455044612216884, + "loss": 0.1151, + "step": 34090 + }, + { + "epoch": 0.5319148936170213, + "grad_norm": 0.14440476894378662, + "learning_rate": 0.00016454004700401406, + "loss": 0.2386, + "step": 34100 + }, + { + "epoch": 0.532070880389343, + "grad_norm": 2.2777981758117676, + "learning_rate": 0.00016452964788585928, + "loss": 0.2009, + "step": 34110 + }, + { + "epoch": 0.5322268671616647, + "grad_norm": 0.9206979274749756, + "learning_rate": 0.0001645192487677045, + "loss": 0.2745, + "step": 34120 + }, + { + "epoch": 0.5323828539339864, + "grad_norm": 1.6947574615478516, + "learning_rate": 0.00016450884964954972, + "loss": 0.2584, + "step": 34130 + }, + { + "epoch": 0.5325388407063081, + "grad_norm": 0.401444673538208, + "learning_rate": 0.00016449845053139494, + "loss": 0.2218, + "step": 34140 + }, + { + "epoch": 0.5326948274786298, + "grad_norm": 0.08261553198099136, + "learning_rate": 0.00016448805141324016, + "loss": 0.2775, + "step": 34150 + }, + { + "epoch": 0.5328508142509515, + "grad_norm": 0.1017974391579628, + "learning_rate": 0.00016447765229508538, + "loss": 0.2095, + "step": 34160 + }, + { + "epoch": 0.5330068010232732, + "grad_norm": 1.3759571313858032, + "learning_rate": 0.0001644672531769306, + "loss": 0.2643, + "step": 34170 + }, + { + "epoch": 0.5331627877955949, + "grad_norm": 1.2654389142990112, + "learning_rate": 0.00016445685405877582, + "loss": 0.2949, + "step": 34180 + }, + { + "epoch": 0.5333187745679167, + "grad_norm": 1.5481843948364258, + "learning_rate": 0.00016444645494062104, + "loss": 0.1264, + "step": 34190 + }, + { + "epoch": 0.5334747613402383, + "grad_norm": 1.8094528913497925, + "learning_rate": 0.00016443605582246626, + "loss": 0.2727, + "step": 34200 + }, + { + "epoch": 0.5336307481125601, + "grad_norm": 2.224538564682007, + "learning_rate": 0.00016442565670431148, + "loss": 0.3096, + "step": 34210 + }, + { + "epoch": 0.5337867348848817, + "grad_norm": 0.6375226974487305, + "learning_rate": 0.0001644152575861567, + "loss": 0.2251, + "step": 34220 + }, + { + "epoch": 0.5339427216572035, + "grad_norm": 3.727106809616089, + "learning_rate": 0.00016440485846800192, + "loss": 0.4374, + "step": 34230 + }, + { + "epoch": 0.5340987084295252, + "grad_norm": 0.13345426321029663, + "learning_rate": 0.00016439445934984714, + "loss": 0.2011, + "step": 34240 + }, + { + "epoch": 0.5342546952018469, + "grad_norm": 2.1658668518066406, + "learning_rate": 0.00016438406023169236, + "loss": 0.2457, + "step": 34250 + }, + { + "epoch": 0.5344106819741686, + "grad_norm": 19.238407135009766, + "learning_rate": 0.00016437366111353758, + "loss": 0.2756, + "step": 34260 + }, + { + "epoch": 0.5345666687464903, + "grad_norm": 1.0292778015136719, + "learning_rate": 0.0001643632619953828, + "loss": 0.1646, + "step": 34270 + }, + { + "epoch": 0.534722655518812, + "grad_norm": 0.9372987747192383, + "learning_rate": 0.00016435286287722802, + "loss": 0.2762, + "step": 34280 + }, + { + "epoch": 0.5348786422911337, + "grad_norm": 0.3918002247810364, + "learning_rate": 0.00016434246375907324, + "loss": 0.1406, + "step": 34290 + }, + { + "epoch": 0.5350346290634554, + "grad_norm": 1.3518732786178589, + "learning_rate": 0.00016433206464091846, + "loss": 0.2992, + "step": 34300 + }, + { + "epoch": 0.5351906158357771, + "grad_norm": 0.73117595911026, + "learning_rate": 0.00016432166552276368, + "loss": 0.1172, + "step": 34310 + }, + { + "epoch": 0.5353466026080989, + "grad_norm": 0.06655958294868469, + "learning_rate": 0.0001643112664046089, + "loss": 0.1332, + "step": 34320 + }, + { + "epoch": 0.5355025893804205, + "grad_norm": 3.2587168216705322, + "learning_rate": 0.00016430086728645412, + "loss": 0.2052, + "step": 34330 + }, + { + "epoch": 0.5356585761527423, + "grad_norm": 2.6668431758880615, + "learning_rate": 0.00016429046816829934, + "loss": 0.1295, + "step": 34340 + }, + { + "epoch": 0.5358145629250639, + "grad_norm": 1.316080927848816, + "learning_rate": 0.00016428006905014456, + "loss": 0.2072, + "step": 34350 + }, + { + "epoch": 0.5359705496973857, + "grad_norm": 5.144528388977051, + "learning_rate": 0.00016426966993198977, + "loss": 0.1934, + "step": 34360 + }, + { + "epoch": 0.5361265364697073, + "grad_norm": 1.3529599905014038, + "learning_rate": 0.000164259270813835, + "loss": 0.2077, + "step": 34370 + }, + { + "epoch": 0.5362825232420291, + "grad_norm": 0.9132925868034363, + "learning_rate": 0.00016424887169568021, + "loss": 0.2404, + "step": 34380 + }, + { + "epoch": 0.5364385100143508, + "grad_norm": 1.520033597946167, + "learning_rate": 0.00016423847257752543, + "loss": 0.1765, + "step": 34390 + }, + { + "epoch": 0.5365944967866725, + "grad_norm": 2.343975305557251, + "learning_rate": 0.00016422807345937065, + "loss": 0.1959, + "step": 34400 + }, + { + "epoch": 0.5367504835589942, + "grad_norm": 0.5855118632316589, + "learning_rate": 0.00016421767434121587, + "loss": 0.0848, + "step": 34410 + }, + { + "epoch": 0.5369064703313159, + "grad_norm": 1.9108648300170898, + "learning_rate": 0.0001642072752230611, + "loss": 0.2406, + "step": 34420 + }, + { + "epoch": 0.5370624571036376, + "grad_norm": 1.0323792695999146, + "learning_rate": 0.0001641968761049063, + "loss": 0.2298, + "step": 34430 + }, + { + "epoch": 0.5372184438759593, + "grad_norm": 0.17435620725154877, + "learning_rate": 0.00016418647698675153, + "loss": 0.1396, + "step": 34440 + }, + { + "epoch": 0.537374430648281, + "grad_norm": 0.3526577353477478, + "learning_rate": 0.00016417607786859675, + "loss": 0.172, + "step": 34450 + }, + { + "epoch": 0.5375304174206027, + "grad_norm": 2.73685884475708, + "learning_rate": 0.00016416567875044197, + "loss": 0.1167, + "step": 34460 + }, + { + "epoch": 0.5376864041929245, + "grad_norm": 2.765693426132202, + "learning_rate": 0.0001641552796322872, + "loss": 0.1356, + "step": 34470 + }, + { + "epoch": 0.5378423909652461, + "grad_norm": 0.47628021240234375, + "learning_rate": 0.0001641448805141324, + "loss": 0.3104, + "step": 34480 + }, + { + "epoch": 0.5379983777375679, + "grad_norm": 1.1307590007781982, + "learning_rate": 0.00016413448139597763, + "loss": 0.3247, + "step": 34490 + }, + { + "epoch": 0.5381543645098895, + "grad_norm": 0.07136381417512894, + "learning_rate": 0.00016412408227782285, + "loss": 0.319, + "step": 34500 + }, + { + "epoch": 0.5383103512822113, + "grad_norm": 1.7850221395492554, + "learning_rate": 0.00016411368315966807, + "loss": 0.2515, + "step": 34510 + }, + { + "epoch": 0.5384663380545329, + "grad_norm": 1.3870742321014404, + "learning_rate": 0.0001641032840415133, + "loss": 0.256, + "step": 34520 + }, + { + "epoch": 0.5386223248268547, + "grad_norm": 1.1329221725463867, + "learning_rate": 0.0001640928849233585, + "loss": 0.2255, + "step": 34530 + }, + { + "epoch": 0.5387783115991764, + "grad_norm": 0.23236137628555298, + "learning_rate": 0.00016408248580520373, + "loss": 0.4053, + "step": 34540 + }, + { + "epoch": 0.5389342983714981, + "grad_norm": 3.350924253463745, + "learning_rate": 0.00016407208668704895, + "loss": 0.2816, + "step": 34550 + }, + { + "epoch": 0.5390902851438198, + "grad_norm": 1.7459170818328857, + "learning_rate": 0.00016406168756889417, + "loss": 0.0834, + "step": 34560 + }, + { + "epoch": 0.5392462719161415, + "grad_norm": 0.89351487159729, + "learning_rate": 0.0001640512884507394, + "loss": 0.1146, + "step": 34570 + }, + { + "epoch": 0.5394022586884633, + "grad_norm": 0.49410831928253174, + "learning_rate": 0.0001640408893325846, + "loss": 0.2319, + "step": 34580 + }, + { + "epoch": 0.5395582454607849, + "grad_norm": 0.009971237741410732, + "learning_rate": 0.00016403049021442983, + "loss": 0.1674, + "step": 34590 + }, + { + "epoch": 0.5397142322331067, + "grad_norm": 1.420511245727539, + "learning_rate": 0.00016402009109627505, + "loss": 0.1469, + "step": 34600 + }, + { + "epoch": 0.5398702190054283, + "grad_norm": 2.5686769485473633, + "learning_rate": 0.00016400969197812027, + "loss": 0.2607, + "step": 34610 + }, + { + "epoch": 0.5400262057777501, + "grad_norm": 2.5808913707733154, + "learning_rate": 0.00016399929285996549, + "loss": 0.4412, + "step": 34620 + }, + { + "epoch": 0.5401821925500717, + "grad_norm": 0.2778591811656952, + "learning_rate": 0.0001639888937418107, + "loss": 0.1563, + "step": 34630 + }, + { + "epoch": 0.5403381793223935, + "grad_norm": 1.6448099613189697, + "learning_rate": 0.00016397849462365592, + "loss": 0.2484, + "step": 34640 + }, + { + "epoch": 0.5404941660947151, + "grad_norm": 0.003213417250663042, + "learning_rate": 0.00016396809550550114, + "loss": 0.1582, + "step": 34650 + }, + { + "epoch": 0.5406501528670369, + "grad_norm": 2.3433990478515625, + "learning_rate": 0.00016395769638734636, + "loss": 0.2774, + "step": 34660 + }, + { + "epoch": 0.5408061396393585, + "grad_norm": 0.09544426947832108, + "learning_rate": 0.00016394729726919158, + "loss": 0.1242, + "step": 34670 + }, + { + "epoch": 0.5409621264116803, + "grad_norm": 1.8934235572814941, + "learning_rate": 0.0001639368981510368, + "loss": 0.1636, + "step": 34680 + }, + { + "epoch": 0.541118113184002, + "grad_norm": 2.8528554439544678, + "learning_rate": 0.00016392649903288202, + "loss": 0.2524, + "step": 34690 + }, + { + "epoch": 0.5412740999563237, + "grad_norm": 2.837836980819702, + "learning_rate": 0.00016391609991472722, + "loss": 0.1808, + "step": 34700 + }, + { + "epoch": 0.5414300867286455, + "grad_norm": 4.349740028381348, + "learning_rate": 0.00016390570079657246, + "loss": 0.12, + "step": 34710 + }, + { + "epoch": 0.5415860735009671, + "grad_norm": 0.6011945605278015, + "learning_rate": 0.00016389530167841765, + "loss": 0.2865, + "step": 34720 + }, + { + "epoch": 0.5417420602732889, + "grad_norm": 1.2079488039016724, + "learning_rate": 0.0001638849025602629, + "loss": 0.125, + "step": 34730 + }, + { + "epoch": 0.5418980470456105, + "grad_norm": 1.9549164772033691, + "learning_rate": 0.0001638745034421081, + "loss": 0.3285, + "step": 34740 + }, + { + "epoch": 0.5420540338179323, + "grad_norm": 0.6969407200813293, + "learning_rate": 0.00016386410432395334, + "loss": 0.3267, + "step": 34750 + }, + { + "epoch": 0.5422100205902539, + "grad_norm": 1.0561965703964233, + "learning_rate": 0.00016385370520579853, + "loss": 0.1463, + "step": 34760 + }, + { + "epoch": 0.5423660073625757, + "grad_norm": 1.8208931684494019, + "learning_rate": 0.00016384330608764378, + "loss": 0.1713, + "step": 34770 + }, + { + "epoch": 0.5425219941348973, + "grad_norm": 0.6349910497665405, + "learning_rate": 0.000163832906969489, + "loss": 0.2244, + "step": 34780 + }, + { + "epoch": 0.5426779809072191, + "grad_norm": 1.1976515054702759, + "learning_rate": 0.00016382250785133422, + "loss": 0.3407, + "step": 34790 + }, + { + "epoch": 0.5428339676795407, + "grad_norm": 0.5970319509506226, + "learning_rate": 0.00016381210873317944, + "loss": 0.1419, + "step": 34800 + }, + { + "epoch": 0.5429899544518625, + "grad_norm": 0.14623159170150757, + "learning_rate": 0.00016380170961502466, + "loss": 0.14, + "step": 34810 + }, + { + "epoch": 0.5431459412241841, + "grad_norm": 1.4332351684570312, + "learning_rate": 0.00016379131049686988, + "loss": 0.1093, + "step": 34820 + }, + { + "epoch": 0.5433019279965059, + "grad_norm": 1.1443063020706177, + "learning_rate": 0.0001637809113787151, + "loss": 0.3676, + "step": 34830 + }, + { + "epoch": 0.5434579147688277, + "grad_norm": 1.3897461891174316, + "learning_rate": 0.00016377051226056032, + "loss": 0.3713, + "step": 34840 + }, + { + "epoch": 0.5436139015411493, + "grad_norm": 1.7773199081420898, + "learning_rate": 0.00016376011314240554, + "loss": 0.2249, + "step": 34850 + }, + { + "epoch": 0.5437698883134711, + "grad_norm": 0.8524389266967773, + "learning_rate": 0.00016374971402425076, + "loss": 0.2158, + "step": 34860 + }, + { + "epoch": 0.5439258750857927, + "grad_norm": 0.5572389960289001, + "learning_rate": 0.00016373931490609598, + "loss": 0.215, + "step": 34870 + }, + { + "epoch": 0.5440818618581145, + "grad_norm": 3.0238709449768066, + "learning_rate": 0.0001637289157879412, + "loss": 0.3765, + "step": 34880 + }, + { + "epoch": 0.5442378486304361, + "grad_norm": 1.1655421257019043, + "learning_rate": 0.00016371851666978642, + "loss": 0.2491, + "step": 34890 + }, + { + "epoch": 0.5443938354027579, + "grad_norm": 0.49492090940475464, + "learning_rate": 0.00016370811755163164, + "loss": 0.1287, + "step": 34900 + }, + { + "epoch": 0.5445498221750795, + "grad_norm": 0.2732921540737152, + "learning_rate": 0.00016369771843347685, + "loss": 0.2154, + "step": 34910 + }, + { + "epoch": 0.5447058089474013, + "grad_norm": 2.5807769298553467, + "learning_rate": 0.00016368731931532207, + "loss": 0.2032, + "step": 34920 + }, + { + "epoch": 0.5448617957197229, + "grad_norm": 0.5044315457344055, + "learning_rate": 0.0001636769201971673, + "loss": 0.2595, + "step": 34930 + }, + { + "epoch": 0.5450177824920447, + "grad_norm": 2.1248972415924072, + "learning_rate": 0.00016366652107901251, + "loss": 0.5651, + "step": 34940 + }, + { + "epoch": 0.5451737692643663, + "grad_norm": 0.8391468524932861, + "learning_rate": 0.00016365612196085773, + "loss": 0.3028, + "step": 34950 + }, + { + "epoch": 0.5453297560366881, + "grad_norm": 0.36081477999687195, + "learning_rate": 0.00016364572284270295, + "loss": 0.1073, + "step": 34960 + }, + { + "epoch": 0.5454857428090097, + "grad_norm": 3.175804376602173, + "learning_rate": 0.00016363532372454817, + "loss": 0.1889, + "step": 34970 + }, + { + "epoch": 0.5456417295813315, + "grad_norm": 0.20647937059402466, + "learning_rate": 0.0001636249246063934, + "loss": 0.2416, + "step": 34980 + }, + { + "epoch": 0.5457977163536533, + "grad_norm": 2.3525943756103516, + "learning_rate": 0.0001636145254882386, + "loss": 0.3179, + "step": 34990 + }, + { + "epoch": 0.5459537031259749, + "grad_norm": 0.2641216814517975, + "learning_rate": 0.00016360412637008383, + "loss": 0.2157, + "step": 35000 + }, + { + "epoch": 0.5461096898982967, + "grad_norm": 0.09906073659658432, + "learning_rate": 0.00016359372725192905, + "loss": 0.2576, + "step": 35010 + }, + { + "epoch": 0.5462656766706183, + "grad_norm": 0.13971920311450958, + "learning_rate": 0.00016358332813377427, + "loss": 0.1298, + "step": 35020 + }, + { + "epoch": 0.5464216634429401, + "grad_norm": 1.4535638093948364, + "learning_rate": 0.0001635729290156195, + "loss": 0.2357, + "step": 35030 + }, + { + "epoch": 0.5465776502152617, + "grad_norm": 0.9107828736305237, + "learning_rate": 0.0001635625298974647, + "loss": 0.3049, + "step": 35040 + }, + { + "epoch": 0.5467336369875835, + "grad_norm": 0.0377386212348938, + "learning_rate": 0.00016355213077930993, + "loss": 0.2066, + "step": 35050 + }, + { + "epoch": 0.5468896237599051, + "grad_norm": 2.7800869941711426, + "learning_rate": 0.00016354173166115515, + "loss": 0.2974, + "step": 35060 + }, + { + "epoch": 0.5470456105322269, + "grad_norm": 1.6247998476028442, + "learning_rate": 0.00016353133254300037, + "loss": 0.3822, + "step": 35070 + }, + { + "epoch": 0.5472015973045485, + "grad_norm": 1.6479015350341797, + "learning_rate": 0.0001635209334248456, + "loss": 0.3401, + "step": 35080 + }, + { + "epoch": 0.5473575840768703, + "grad_norm": 0.1879737675189972, + "learning_rate": 0.0001635105343066908, + "loss": 0.1549, + "step": 35090 + }, + { + "epoch": 0.5475135708491919, + "grad_norm": 0.39355361461639404, + "learning_rate": 0.00016350013518853603, + "loss": 0.2032, + "step": 35100 + }, + { + "epoch": 0.5476695576215137, + "grad_norm": 2.9976983070373535, + "learning_rate": 0.00016348973607038125, + "loss": 0.3867, + "step": 35110 + }, + { + "epoch": 0.5478255443938354, + "grad_norm": 0.8242707252502441, + "learning_rate": 0.00016347933695222647, + "loss": 0.3601, + "step": 35120 + }, + { + "epoch": 0.5479815311661571, + "grad_norm": 0.874955415725708, + "learning_rate": 0.0001634689378340717, + "loss": 0.2709, + "step": 35130 + }, + { + "epoch": 0.5481375179384789, + "grad_norm": 0.05980971083045006, + "learning_rate": 0.0001634585387159169, + "loss": 0.2646, + "step": 35140 + }, + { + "epoch": 0.5482935047108005, + "grad_norm": 1.8437328338623047, + "learning_rate": 0.0001634481395977621, + "loss": 0.1965, + "step": 35150 + }, + { + "epoch": 0.5484494914831223, + "grad_norm": 0.4142405688762665, + "learning_rate": 0.00016343774047960735, + "loss": 0.1756, + "step": 35160 + }, + { + "epoch": 0.5486054782554439, + "grad_norm": 4.836324214935303, + "learning_rate": 0.00016342734136145254, + "loss": 0.2534, + "step": 35170 + }, + { + "epoch": 0.5487614650277657, + "grad_norm": 1.4933065176010132, + "learning_rate": 0.00016341694224329779, + "loss": 0.2355, + "step": 35180 + }, + { + "epoch": 0.5489174518000873, + "grad_norm": 1.9468894004821777, + "learning_rate": 0.00016340654312514298, + "loss": 0.3113, + "step": 35190 + }, + { + "epoch": 0.5490734385724091, + "grad_norm": 1.012710452079773, + "learning_rate": 0.00016339614400698822, + "loss": 0.2635, + "step": 35200 + }, + { + "epoch": 0.5492294253447307, + "grad_norm": 1.5551140308380127, + "learning_rate": 0.00016338574488883342, + "loss": 0.2951, + "step": 35210 + }, + { + "epoch": 0.5493854121170525, + "grad_norm": 1.8613696098327637, + "learning_rate": 0.00016337534577067866, + "loss": 0.2832, + "step": 35220 + }, + { + "epoch": 0.5495413988893741, + "grad_norm": 1.6748839616775513, + "learning_rate": 0.00016336494665252386, + "loss": 0.2266, + "step": 35230 + }, + { + "epoch": 0.5496973856616959, + "grad_norm": 0.4094032347202301, + "learning_rate": 0.0001633545475343691, + "loss": 0.3497, + "step": 35240 + }, + { + "epoch": 0.5498533724340176, + "grad_norm": 1.6414631605148315, + "learning_rate": 0.0001633441484162143, + "loss": 0.1763, + "step": 35250 + }, + { + "epoch": 0.5500093592063393, + "grad_norm": 1.6380645036697388, + "learning_rate": 0.00016333374929805954, + "loss": 0.1355, + "step": 35260 + }, + { + "epoch": 0.550165345978661, + "grad_norm": 0.7325630187988281, + "learning_rate": 0.00016332335017990473, + "loss": 0.3105, + "step": 35270 + }, + { + "epoch": 0.5503213327509827, + "grad_norm": 1.6015644073486328, + "learning_rate": 0.00016331295106174998, + "loss": 0.2311, + "step": 35280 + }, + { + "epoch": 0.5504773195233045, + "grad_norm": 0.9224210381507874, + "learning_rate": 0.00016330255194359517, + "loss": 0.2746, + "step": 35290 + }, + { + "epoch": 0.5506333062956261, + "grad_norm": 1.30025315284729, + "learning_rate": 0.00016329215282544042, + "loss": 0.2066, + "step": 35300 + }, + { + "epoch": 0.5507892930679479, + "grad_norm": 0.481125146150589, + "learning_rate": 0.0001632817537072856, + "loss": 0.1778, + "step": 35310 + }, + { + "epoch": 0.5509452798402695, + "grad_norm": 0.2709486782550812, + "learning_rate": 0.00016327135458913086, + "loss": 0.1131, + "step": 35320 + }, + { + "epoch": 0.5511012666125913, + "grad_norm": 0.6248563528060913, + "learning_rate": 0.00016326095547097605, + "loss": 0.4356, + "step": 35330 + }, + { + "epoch": 0.5512572533849129, + "grad_norm": 3.0947647094726562, + "learning_rate": 0.0001632505563528213, + "loss": 0.3072, + "step": 35340 + }, + { + "epoch": 0.5514132401572347, + "grad_norm": 1.009535312652588, + "learning_rate": 0.0001632401572346665, + "loss": 0.1808, + "step": 35350 + }, + { + "epoch": 0.5515692269295563, + "grad_norm": 0.042604975402355194, + "learning_rate": 0.00016322975811651174, + "loss": 0.1198, + "step": 35360 + }, + { + "epoch": 0.5517252137018781, + "grad_norm": 1.8652396202087402, + "learning_rate": 0.00016321935899835693, + "loss": 0.2176, + "step": 35370 + }, + { + "epoch": 0.5518812004741998, + "grad_norm": 4.6097187995910645, + "learning_rate": 0.00016320895988020218, + "loss": 0.2132, + "step": 35380 + }, + { + "epoch": 0.5520371872465215, + "grad_norm": 1.3615522384643555, + "learning_rate": 0.00016319856076204737, + "loss": 0.2751, + "step": 35390 + }, + { + "epoch": 0.5521931740188432, + "grad_norm": 0.8844773769378662, + "learning_rate": 0.00016318816164389262, + "loss": 0.22, + "step": 35400 + }, + { + "epoch": 0.5523491607911649, + "grad_norm": 0.5413331985473633, + "learning_rate": 0.0001631777625257378, + "loss": 0.2159, + "step": 35410 + }, + { + "epoch": 0.5525051475634866, + "grad_norm": 1.5456678867340088, + "learning_rate": 0.00016316736340758306, + "loss": 0.2244, + "step": 35420 + }, + { + "epoch": 0.5526611343358083, + "grad_norm": 2.1405861377716064, + "learning_rate": 0.00016315696428942825, + "loss": 0.446, + "step": 35430 + }, + { + "epoch": 0.5528171211081301, + "grad_norm": 1.4269858598709106, + "learning_rate": 0.0001631465651712735, + "loss": 0.3599, + "step": 35440 + }, + { + "epoch": 0.5529731078804517, + "grad_norm": 1.8682516813278198, + "learning_rate": 0.0001631361660531187, + "loss": 0.115, + "step": 35450 + }, + { + "epoch": 0.5531290946527735, + "grad_norm": 1.0175774097442627, + "learning_rate": 0.00016312576693496394, + "loss": 0.4003, + "step": 35460 + }, + { + "epoch": 0.5532850814250951, + "grad_norm": 2.4369170665740967, + "learning_rate": 0.00016311536781680913, + "loss": 0.2489, + "step": 35470 + }, + { + "epoch": 0.5534410681974169, + "grad_norm": 1.886022686958313, + "learning_rate": 0.00016310496869865437, + "loss": 0.4201, + "step": 35480 + }, + { + "epoch": 0.5535970549697385, + "grad_norm": 1.5811959505081177, + "learning_rate": 0.00016309456958049957, + "loss": 0.4286, + "step": 35490 + }, + { + "epoch": 0.5537530417420603, + "grad_norm": 1.971110224723816, + "learning_rate": 0.00016308417046234481, + "loss": 0.3599, + "step": 35500 + }, + { + "epoch": 0.553909028514382, + "grad_norm": 0.0335380844771862, + "learning_rate": 0.00016307377134419, + "loss": 0.3357, + "step": 35510 + }, + { + "epoch": 0.5540650152867037, + "grad_norm": 1.8519577980041504, + "learning_rate": 0.00016306337222603525, + "loss": 0.3126, + "step": 35520 + }, + { + "epoch": 0.5542210020590254, + "grad_norm": 2.1563379764556885, + "learning_rate": 0.00016305297310788045, + "loss": 0.1123, + "step": 35530 + }, + { + "epoch": 0.5543769888313471, + "grad_norm": 1.64332914352417, + "learning_rate": 0.0001630425739897257, + "loss": 0.2853, + "step": 35540 + }, + { + "epoch": 0.5545329756036688, + "grad_norm": 0.061150554567575455, + "learning_rate": 0.00016303217487157088, + "loss": 0.1943, + "step": 35550 + }, + { + "epoch": 0.5546889623759905, + "grad_norm": 1.2701060771942139, + "learning_rate": 0.00016302177575341613, + "loss": 0.2304, + "step": 35560 + }, + { + "epoch": 0.5548449491483122, + "grad_norm": 2.424860715866089, + "learning_rate": 0.00016301137663526132, + "loss": 0.2128, + "step": 35570 + }, + { + "epoch": 0.5550009359206339, + "grad_norm": 0.6803575158119202, + "learning_rate": 0.00016300097751710657, + "loss": 0.1145, + "step": 35580 + }, + { + "epoch": 0.5551569226929557, + "grad_norm": 1.2855092287063599, + "learning_rate": 0.00016299057839895176, + "loss": 0.2448, + "step": 35590 + }, + { + "epoch": 0.5553129094652773, + "grad_norm": 2.6340911388397217, + "learning_rate": 0.00016298017928079698, + "loss": 0.1856, + "step": 35600 + }, + { + "epoch": 0.5554688962375991, + "grad_norm": 1.776382327079773, + "learning_rate": 0.0001629697801626422, + "loss": 0.2422, + "step": 35610 + }, + { + "epoch": 0.5556248830099207, + "grad_norm": 3.0746347904205322, + "learning_rate": 0.00016295938104448742, + "loss": 0.1578, + "step": 35620 + }, + { + "epoch": 0.5557808697822425, + "grad_norm": 1.4887659549713135, + "learning_rate": 0.00016294898192633267, + "loss": 0.2385, + "step": 35630 + }, + { + "epoch": 0.5559368565545642, + "grad_norm": 0.5371220707893372, + "learning_rate": 0.00016293858280817786, + "loss": 0.2361, + "step": 35640 + }, + { + "epoch": 0.5560928433268859, + "grad_norm": 0.18604904413223267, + "learning_rate": 0.0001629281836900231, + "loss": 0.0828, + "step": 35650 + }, + { + "epoch": 0.5562488300992076, + "grad_norm": 2.267854690551758, + "learning_rate": 0.0001629177845718683, + "loss": 0.413, + "step": 35660 + }, + { + "epoch": 0.5564048168715293, + "grad_norm": 1.9755452871322632, + "learning_rate": 0.00016290738545371355, + "loss": 0.146, + "step": 35670 + }, + { + "epoch": 0.556560803643851, + "grad_norm": 2.430293321609497, + "learning_rate": 0.00016289698633555874, + "loss": 0.2677, + "step": 35680 + }, + { + "epoch": 0.5567167904161727, + "grad_norm": 0.6319543719291687, + "learning_rate": 0.000162886587217404, + "loss": 0.1481, + "step": 35690 + }, + { + "epoch": 0.5568727771884944, + "grad_norm": 0.17439277470111847, + "learning_rate": 0.00016287618809924918, + "loss": 0.1641, + "step": 35700 + }, + { + "epoch": 0.5570287639608161, + "grad_norm": 0.350175142288208, + "learning_rate": 0.00016286578898109443, + "loss": 0.393, + "step": 35710 + }, + { + "epoch": 0.5571847507331378, + "grad_norm": 2.641941547393799, + "learning_rate": 0.00016285538986293962, + "loss": 0.1907, + "step": 35720 + }, + { + "epoch": 0.5573407375054595, + "grad_norm": 0.1910303384065628, + "learning_rate": 0.00016284499074478487, + "loss": 0.2248, + "step": 35730 + }, + { + "epoch": 0.5574967242777813, + "grad_norm": 1.0180896520614624, + "learning_rate": 0.00016283459162663006, + "loss": 0.1924, + "step": 35740 + }, + { + "epoch": 0.557652711050103, + "grad_norm": 1.2047260999679565, + "learning_rate": 0.0001628241925084753, + "loss": 0.2031, + "step": 35750 + }, + { + "epoch": 0.5578086978224247, + "grad_norm": 1.8182405233383179, + "learning_rate": 0.0001628137933903205, + "loss": 0.1573, + "step": 35760 + }, + { + "epoch": 0.5579646845947464, + "grad_norm": 0.5485963821411133, + "learning_rate": 0.00016280339427216574, + "loss": 0.152, + "step": 35770 + }, + { + "epoch": 0.5581206713670681, + "grad_norm": 0.1674145758152008, + "learning_rate": 0.00016279299515401094, + "loss": 0.2388, + "step": 35780 + }, + { + "epoch": 0.5582766581393898, + "grad_norm": 0.36187657713890076, + "learning_rate": 0.00016278259603585618, + "loss": 0.1618, + "step": 35790 + }, + { + "epoch": 0.5584326449117115, + "grad_norm": 0.013768521137535572, + "learning_rate": 0.00016277219691770138, + "loss": 0.3666, + "step": 35800 + }, + { + "epoch": 0.5585886316840332, + "grad_norm": 2.8602802753448486, + "learning_rate": 0.00016276179779954662, + "loss": 0.2065, + "step": 35810 + }, + { + "epoch": 0.5587446184563549, + "grad_norm": 1.7925455570220947, + "learning_rate": 0.00016275139868139182, + "loss": 0.1729, + "step": 35820 + }, + { + "epoch": 0.5589006052286766, + "grad_norm": 2.910456418991089, + "learning_rate": 0.00016274099956323706, + "loss": 0.3734, + "step": 35830 + }, + { + "epoch": 0.5590565920009983, + "grad_norm": 0.8378308415412903, + "learning_rate": 0.00016273060044508225, + "loss": 0.2255, + "step": 35840 + }, + { + "epoch": 0.55921257877332, + "grad_norm": 0.409534752368927, + "learning_rate": 0.0001627202013269275, + "loss": 0.2046, + "step": 35850 + }, + { + "epoch": 0.5593685655456417, + "grad_norm": 0.49498099088668823, + "learning_rate": 0.0001627098022087727, + "loss": 0.1058, + "step": 35860 + }, + { + "epoch": 0.5595245523179634, + "grad_norm": 1.1761783361434937, + "learning_rate": 0.00016269940309061794, + "loss": 0.1566, + "step": 35870 + }, + { + "epoch": 0.5596805390902851, + "grad_norm": 3.8204751014709473, + "learning_rate": 0.00016268900397246313, + "loss": 0.2646, + "step": 35880 + }, + { + "epoch": 0.5598365258626069, + "grad_norm": 0.9882522225379944, + "learning_rate": 0.00016267860485430838, + "loss": 0.1756, + "step": 35890 + }, + { + "epoch": 0.5599925126349286, + "grad_norm": 1.1832259893417358, + "learning_rate": 0.00016266820573615357, + "loss": 0.1385, + "step": 35900 + }, + { + "epoch": 0.5601484994072503, + "grad_norm": 0.7638296484947205, + "learning_rate": 0.00016265780661799882, + "loss": 0.5679, + "step": 35910 + }, + { + "epoch": 0.560304486179572, + "grad_norm": 0.8551504611968994, + "learning_rate": 0.000162647407499844, + "loss": 0.2981, + "step": 35920 + }, + { + "epoch": 0.5604604729518937, + "grad_norm": 5.499948501586914, + "learning_rate": 0.00016263700838168926, + "loss": 0.1429, + "step": 35930 + }, + { + "epoch": 0.5606164597242154, + "grad_norm": 3.0494496822357178, + "learning_rate": 0.00016262660926353445, + "loss": 0.1322, + "step": 35940 + }, + { + "epoch": 0.5607724464965371, + "grad_norm": 0.2482384592294693, + "learning_rate": 0.0001626162101453797, + "loss": 0.2473, + "step": 35950 + }, + { + "epoch": 0.5609284332688588, + "grad_norm": 0.009750776924192905, + "learning_rate": 0.0001626058110272249, + "loss": 0.0429, + "step": 35960 + }, + { + "epoch": 0.5610844200411805, + "grad_norm": 0.5742604732513428, + "learning_rate": 0.00016259541190907014, + "loss": 0.4565, + "step": 35970 + }, + { + "epoch": 0.5612404068135022, + "grad_norm": 0.7810243368148804, + "learning_rate": 0.00016258501279091533, + "loss": 0.2107, + "step": 35980 + }, + { + "epoch": 0.5613963935858239, + "grad_norm": 2.785747528076172, + "learning_rate": 0.00016257461367276058, + "loss": 0.5087, + "step": 35990 + }, + { + "epoch": 0.5615523803581456, + "grad_norm": 0.8806902766227722, + "learning_rate": 0.00016256421455460577, + "loss": 0.2937, + "step": 36000 + }, + { + "epoch": 0.5617083671304673, + "grad_norm": 1.857373833656311, + "learning_rate": 0.00016255381543645102, + "loss": 0.5042, + "step": 36010 + }, + { + "epoch": 0.561864353902789, + "grad_norm": 0.5333901047706604, + "learning_rate": 0.0001625434163182962, + "loss": 0.1678, + "step": 36020 + }, + { + "epoch": 0.5620203406751108, + "grad_norm": 11.023160934448242, + "learning_rate": 0.00016253301720014145, + "loss": 0.2721, + "step": 36030 + }, + { + "epoch": 0.5621763274474325, + "grad_norm": 2.6214029788970947, + "learning_rate": 0.00016252261808198665, + "loss": 0.1675, + "step": 36040 + }, + { + "epoch": 0.5623323142197542, + "grad_norm": 0.25980064272880554, + "learning_rate": 0.0001625122189638319, + "loss": 0.1832, + "step": 36050 + }, + { + "epoch": 0.5624883009920759, + "grad_norm": 1.3559473752975464, + "learning_rate": 0.0001625018198456771, + "loss": 0.121, + "step": 36060 + }, + { + "epoch": 0.5626442877643976, + "grad_norm": 0.3558153212070465, + "learning_rate": 0.0001624914207275223, + "loss": 0.2797, + "step": 36070 + }, + { + "epoch": 0.5628002745367193, + "grad_norm": 0.6314427256584167, + "learning_rate": 0.00016248102160936753, + "loss": 0.3455, + "step": 36080 + }, + { + "epoch": 0.562956261309041, + "grad_norm": 0.16120101511478424, + "learning_rate": 0.00016247062249121275, + "loss": 0.3669, + "step": 36090 + }, + { + "epoch": 0.5631122480813627, + "grad_norm": 1.041572093963623, + "learning_rate": 0.00016246022337305797, + "loss": 0.2744, + "step": 36100 + }, + { + "epoch": 0.5632682348536844, + "grad_norm": 0.2645890414714813, + "learning_rate": 0.00016244982425490318, + "loss": 0.1487, + "step": 36110 + }, + { + "epoch": 0.5634242216260061, + "grad_norm": 5.698698043823242, + "learning_rate": 0.0001624394251367484, + "loss": 0.173, + "step": 36120 + }, + { + "epoch": 0.5635802083983278, + "grad_norm": 0.3299804627895355, + "learning_rate": 0.00016242902601859362, + "loss": 0.1965, + "step": 36130 + }, + { + "epoch": 0.5637361951706495, + "grad_norm": 1.2476481199264526, + "learning_rate": 0.00016241862690043884, + "loss": 0.4097, + "step": 36140 + }, + { + "epoch": 0.5638921819429712, + "grad_norm": 0.30926263332366943, + "learning_rate": 0.00016240822778228406, + "loss": 0.1416, + "step": 36150 + }, + { + "epoch": 0.564048168715293, + "grad_norm": 0.05872740224003792, + "learning_rate": 0.00016239782866412928, + "loss": 0.1449, + "step": 36160 + }, + { + "epoch": 0.5642041554876146, + "grad_norm": 0.2292211800813675, + "learning_rate": 0.0001623874295459745, + "loss": 0.2557, + "step": 36170 + }, + { + "epoch": 0.5643601422599364, + "grad_norm": 1.7822531461715698, + "learning_rate": 0.00016237703042781972, + "loss": 0.2953, + "step": 36180 + }, + { + "epoch": 0.5645161290322581, + "grad_norm": 2.4908461570739746, + "learning_rate": 0.00016236663130966494, + "loss": 0.3888, + "step": 36190 + }, + { + "epoch": 0.5646721158045798, + "grad_norm": 0.08363594114780426, + "learning_rate": 0.00016235623219151016, + "loss": 0.2293, + "step": 36200 + }, + { + "epoch": 0.5648281025769015, + "grad_norm": 1.4830002784729004, + "learning_rate": 0.00016234583307335538, + "loss": 0.2348, + "step": 36210 + }, + { + "epoch": 0.5649840893492232, + "grad_norm": 1.8443071842193604, + "learning_rate": 0.0001623354339552006, + "loss": 0.3392, + "step": 36220 + }, + { + "epoch": 0.5651400761215449, + "grad_norm": 2.8050875663757324, + "learning_rate": 0.00016232503483704582, + "loss": 0.2426, + "step": 36230 + }, + { + "epoch": 0.5652960628938666, + "grad_norm": 3.6627259254455566, + "learning_rate": 0.00016231463571889104, + "loss": 0.2118, + "step": 36240 + }, + { + "epoch": 0.5654520496661883, + "grad_norm": 0.3443094789981842, + "learning_rate": 0.00016230423660073626, + "loss": 0.2418, + "step": 36250 + }, + { + "epoch": 0.56560803643851, + "grad_norm": 1.7411112785339355, + "learning_rate": 0.00016229383748258148, + "loss": 0.1628, + "step": 36260 + }, + { + "epoch": 0.5657640232108317, + "grad_norm": 1.1321816444396973, + "learning_rate": 0.0001622834383644267, + "loss": 0.2732, + "step": 36270 + }, + { + "epoch": 0.5659200099831534, + "grad_norm": 2.9686615467071533, + "learning_rate": 0.00016227303924627192, + "loss": 0.3036, + "step": 36280 + }, + { + "epoch": 0.5660759967554752, + "grad_norm": 5.03530740737915, + "learning_rate": 0.00016226264012811714, + "loss": 0.2143, + "step": 36290 + }, + { + "epoch": 0.5662319835277968, + "grad_norm": 2.3943281173706055, + "learning_rate": 0.00016225224100996236, + "loss": 0.2629, + "step": 36300 + }, + { + "epoch": 0.5663879703001186, + "grad_norm": 1.812828779220581, + "learning_rate": 0.00016224184189180758, + "loss": 0.3892, + "step": 36310 + }, + { + "epoch": 0.5665439570724402, + "grad_norm": 1.4708483219146729, + "learning_rate": 0.0001622314427736528, + "loss": 0.2118, + "step": 36320 + }, + { + "epoch": 0.566699943844762, + "grad_norm": 8.99713134765625, + "learning_rate": 0.00016222104365549802, + "loss": 0.3185, + "step": 36330 + }, + { + "epoch": 0.5668559306170837, + "grad_norm": 1.7472341060638428, + "learning_rate": 0.00016221064453734324, + "loss": 0.1659, + "step": 36340 + }, + { + "epoch": 0.5670119173894054, + "grad_norm": 0.8059778213500977, + "learning_rate": 0.00016220024541918846, + "loss": 0.0621, + "step": 36350 + }, + { + "epoch": 0.5671679041617271, + "grad_norm": 4.0501861572265625, + "learning_rate": 0.00016218984630103368, + "loss": 0.1399, + "step": 36360 + }, + { + "epoch": 0.5673238909340488, + "grad_norm": 0.014400321058928967, + "learning_rate": 0.0001621794471828789, + "loss": 0.2372, + "step": 36370 + }, + { + "epoch": 0.5674798777063705, + "grad_norm": 0.4061933755874634, + "learning_rate": 0.00016216904806472412, + "loss": 0.3854, + "step": 36380 + }, + { + "epoch": 0.5676358644786922, + "grad_norm": 0.4940955340862274, + "learning_rate": 0.00016215864894656933, + "loss": 0.1689, + "step": 36390 + }, + { + "epoch": 0.567791851251014, + "grad_norm": 0.3220517635345459, + "learning_rate": 0.00016214824982841455, + "loss": 0.1912, + "step": 36400 + }, + { + "epoch": 0.5679478380233356, + "grad_norm": 2.0848920345306396, + "learning_rate": 0.00016213785071025977, + "loss": 0.1639, + "step": 36410 + }, + { + "epoch": 0.5681038247956574, + "grad_norm": 0.04129406809806824, + "learning_rate": 0.000162127451592105, + "loss": 0.1378, + "step": 36420 + }, + { + "epoch": 0.568259811567979, + "grad_norm": 1.9343401193618774, + "learning_rate": 0.0001621170524739502, + "loss": 0.3759, + "step": 36430 + }, + { + "epoch": 0.5684157983403008, + "grad_norm": 1.0625615119934082, + "learning_rate": 0.00016210665335579543, + "loss": 0.2074, + "step": 36440 + }, + { + "epoch": 0.5685717851126224, + "grad_norm": 0.14784491062164307, + "learning_rate": 0.00016209625423764065, + "loss": 0.0981, + "step": 36450 + }, + { + "epoch": 0.5687277718849442, + "grad_norm": 1.8117821216583252, + "learning_rate": 0.00016208585511948587, + "loss": 0.2755, + "step": 36460 + }, + { + "epoch": 0.5688837586572658, + "grad_norm": 0.061010006815195084, + "learning_rate": 0.0001620754560013311, + "loss": 0.1815, + "step": 36470 + }, + { + "epoch": 0.5690397454295876, + "grad_norm": 0.44453224539756775, + "learning_rate": 0.0001620650568831763, + "loss": 0.0891, + "step": 36480 + }, + { + "epoch": 0.5691957322019093, + "grad_norm": 0.6130645871162415, + "learning_rate": 0.00016205465776502153, + "loss": 0.2778, + "step": 36490 + }, + { + "epoch": 0.569351718974231, + "grad_norm": 5.408796787261963, + "learning_rate": 0.00016204425864686678, + "loss": 0.2301, + "step": 36500 + }, + { + "epoch": 0.5695077057465527, + "grad_norm": 1.2095412015914917, + "learning_rate": 0.00016203385952871197, + "loss": 0.1378, + "step": 36510 + }, + { + "epoch": 0.5696636925188744, + "grad_norm": 0.46873530745506287, + "learning_rate": 0.0001620234604105572, + "loss": 0.1431, + "step": 36520 + }, + { + "epoch": 0.5698196792911961, + "grad_norm": 1.028438925743103, + "learning_rate": 0.0001620130612924024, + "loss": 0.1114, + "step": 36530 + }, + { + "epoch": 0.5699756660635178, + "grad_norm": 0.5952587127685547, + "learning_rate": 0.00016200266217424763, + "loss": 0.2156, + "step": 36540 + }, + { + "epoch": 0.5701316528358396, + "grad_norm": 0.5134285688400269, + "learning_rate": 0.00016199226305609285, + "loss": 0.2712, + "step": 36550 + }, + { + "epoch": 0.5702876396081612, + "grad_norm": 0.03440163657069206, + "learning_rate": 0.00016198186393793807, + "loss": 0.312, + "step": 36560 + }, + { + "epoch": 0.570443626380483, + "grad_norm": 0.10363951325416565, + "learning_rate": 0.0001619714648197833, + "loss": 0.2399, + "step": 36570 + }, + { + "epoch": 0.5705996131528046, + "grad_norm": 1.3341947793960571, + "learning_rate": 0.0001619610657016285, + "loss": 0.1397, + "step": 36580 + }, + { + "epoch": 0.5707555999251264, + "grad_norm": 0.0668073296546936, + "learning_rate": 0.00016195066658347373, + "loss": 0.1091, + "step": 36590 + }, + { + "epoch": 0.570911586697448, + "grad_norm": 5.0486674308776855, + "learning_rate": 0.00016194026746531895, + "loss": 0.4893, + "step": 36600 + }, + { + "epoch": 0.5710675734697698, + "grad_norm": 1.2655692100524902, + "learning_rate": 0.00016192986834716417, + "loss": 0.3344, + "step": 36610 + }, + { + "epoch": 0.5712235602420914, + "grad_norm": 0.1458190381526947, + "learning_rate": 0.00016191946922900939, + "loss": 0.247, + "step": 36620 + }, + { + "epoch": 0.5713795470144132, + "grad_norm": 2.5789072513580322, + "learning_rate": 0.0001619090701108546, + "loss": 0.1546, + "step": 36630 + }, + { + "epoch": 0.5715355337867349, + "grad_norm": 1.191821575164795, + "learning_rate": 0.00016189867099269983, + "loss": 0.2215, + "step": 36640 + }, + { + "epoch": 0.5716915205590566, + "grad_norm": 8.163617134094238, + "learning_rate": 0.00016188827187454505, + "loss": 0.1696, + "step": 36650 + }, + { + "epoch": 0.5718475073313783, + "grad_norm": 3.162277936935425, + "learning_rate": 0.00016187787275639027, + "loss": 0.2906, + "step": 36660 + }, + { + "epoch": 0.5720034941037, + "grad_norm": 2.2368602752685547, + "learning_rate": 0.00016186747363823548, + "loss": 0.3806, + "step": 36670 + }, + { + "epoch": 0.5721594808760218, + "grad_norm": 2.0193803310394287, + "learning_rate": 0.0001618570745200807, + "loss": 0.1964, + "step": 36680 + }, + { + "epoch": 0.5723154676483434, + "grad_norm": 0.9131811261177063, + "learning_rate": 0.00016184667540192592, + "loss": 0.1902, + "step": 36690 + }, + { + "epoch": 0.5724714544206652, + "grad_norm": 0.21330870687961578, + "learning_rate": 0.00016183627628377114, + "loss": 0.2907, + "step": 36700 + }, + { + "epoch": 0.5726274411929868, + "grad_norm": 0.7145895957946777, + "learning_rate": 0.00016182587716561636, + "loss": 0.2574, + "step": 36710 + }, + { + "epoch": 0.5727834279653086, + "grad_norm": 1.0699433088302612, + "learning_rate": 0.00016181547804746158, + "loss": 0.1675, + "step": 36720 + }, + { + "epoch": 0.5729394147376302, + "grad_norm": 0.5066679120063782, + "learning_rate": 0.0001618050789293068, + "loss": 0.4487, + "step": 36730 + }, + { + "epoch": 0.573095401509952, + "grad_norm": 0.7482333183288574, + "learning_rate": 0.00016179467981115202, + "loss": 0.4101, + "step": 36740 + }, + { + "epoch": 0.5732513882822736, + "grad_norm": 0.15925532579421997, + "learning_rate": 0.00016178428069299724, + "loss": 0.2825, + "step": 36750 + }, + { + "epoch": 0.5734073750545954, + "grad_norm": 0.95689457654953, + "learning_rate": 0.00016177388157484246, + "loss": 0.2754, + "step": 36760 + }, + { + "epoch": 0.573563361826917, + "grad_norm": 2.15043568611145, + "learning_rate": 0.00016176348245668768, + "loss": 0.2073, + "step": 36770 + }, + { + "epoch": 0.5737193485992388, + "grad_norm": 2.651520252227783, + "learning_rate": 0.0001617530833385329, + "loss": 0.2517, + "step": 36780 + }, + { + "epoch": 0.5738753353715605, + "grad_norm": 2.1156702041625977, + "learning_rate": 0.00016174268422037812, + "loss": 0.3029, + "step": 36790 + }, + { + "epoch": 0.5740313221438822, + "grad_norm": 3.6200709342956543, + "learning_rate": 0.00016173228510222334, + "loss": 0.4174, + "step": 36800 + }, + { + "epoch": 0.574187308916204, + "grad_norm": 1.447937250137329, + "learning_rate": 0.00016172188598406856, + "loss": 0.1763, + "step": 36810 + }, + { + "epoch": 0.5743432956885256, + "grad_norm": 1.4020819664001465, + "learning_rate": 0.00016171148686591378, + "loss": 0.3994, + "step": 36820 + }, + { + "epoch": 0.5744992824608474, + "grad_norm": 1.4079807996749878, + "learning_rate": 0.000161701087747759, + "loss": 0.2051, + "step": 36830 + }, + { + "epoch": 0.574655269233169, + "grad_norm": 0.05268567427992821, + "learning_rate": 0.00016169068862960422, + "loss": 0.1773, + "step": 36840 + }, + { + "epoch": 0.5748112560054908, + "grad_norm": 2.248767614364624, + "learning_rate": 0.00016168028951144944, + "loss": 0.2534, + "step": 36850 + }, + { + "epoch": 0.5749672427778124, + "grad_norm": 0.3726302981376648, + "learning_rate": 0.00016166989039329466, + "loss": 0.328, + "step": 36860 + }, + { + "epoch": 0.5751232295501342, + "grad_norm": 1.0772110223770142, + "learning_rate": 0.00016165949127513988, + "loss": 0.2069, + "step": 36870 + }, + { + "epoch": 0.5752792163224558, + "grad_norm": 0.11339398473501205, + "learning_rate": 0.0001616490921569851, + "loss": 0.1827, + "step": 36880 + }, + { + "epoch": 0.5754352030947776, + "grad_norm": 1.2981103658676147, + "learning_rate": 0.00016163869303883032, + "loss": 0.1378, + "step": 36890 + }, + { + "epoch": 0.5755911898670992, + "grad_norm": 0.3302285075187683, + "learning_rate": 0.00016162829392067554, + "loss": 0.2311, + "step": 36900 + }, + { + "epoch": 0.575747176639421, + "grad_norm": 0.10168436169624329, + "learning_rate": 0.00016161789480252076, + "loss": 0.1678, + "step": 36910 + }, + { + "epoch": 0.5759031634117426, + "grad_norm": 1.6778357028961182, + "learning_rate": 0.00016160749568436598, + "loss": 0.2385, + "step": 36920 + }, + { + "epoch": 0.5760591501840644, + "grad_norm": 8.1710205078125, + "learning_rate": 0.0001615970965662112, + "loss": 0.3287, + "step": 36930 + }, + { + "epoch": 0.5762151369563862, + "grad_norm": 2.6481053829193115, + "learning_rate": 0.00016158669744805641, + "loss": 0.3119, + "step": 36940 + }, + { + "epoch": 0.5763711237287078, + "grad_norm": 0.3120105266571045, + "learning_rate": 0.00016157629832990163, + "loss": 0.1985, + "step": 36950 + }, + { + "epoch": 0.5765271105010296, + "grad_norm": 2.7221410274505615, + "learning_rate": 0.00016156589921174685, + "loss": 0.3697, + "step": 36960 + }, + { + "epoch": 0.5766830972733512, + "grad_norm": 0.34397706389427185, + "learning_rate": 0.00016155550009359207, + "loss": 0.2162, + "step": 36970 + }, + { + "epoch": 0.576839084045673, + "grad_norm": 4.6421003341674805, + "learning_rate": 0.0001615451009754373, + "loss": 0.2892, + "step": 36980 + }, + { + "epoch": 0.5769950708179946, + "grad_norm": 1.5395855903625488, + "learning_rate": 0.0001615347018572825, + "loss": 0.1258, + "step": 36990 + }, + { + "epoch": 0.5771510575903164, + "grad_norm": 0.4586782455444336, + "learning_rate": 0.00016152430273912773, + "loss": 0.1752, + "step": 37000 + }, + { + "epoch": 0.577307044362638, + "grad_norm": 0.4277658760547638, + "learning_rate": 0.00016151390362097295, + "loss": 0.1722, + "step": 37010 + }, + { + "epoch": 0.5774630311349598, + "grad_norm": 1.0282249450683594, + "learning_rate": 0.00016150350450281817, + "loss": 0.1361, + "step": 37020 + }, + { + "epoch": 0.5776190179072814, + "grad_norm": 0.2496921867132187, + "learning_rate": 0.0001614931053846634, + "loss": 0.0954, + "step": 37030 + }, + { + "epoch": 0.5777750046796032, + "grad_norm": 1.7900562286376953, + "learning_rate": 0.0001614827062665086, + "loss": 0.1284, + "step": 37040 + }, + { + "epoch": 0.5779309914519248, + "grad_norm": 0.6629091501235962, + "learning_rate": 0.00016147230714835383, + "loss": 0.2657, + "step": 37050 + }, + { + "epoch": 0.5780869782242466, + "grad_norm": 2.502882242202759, + "learning_rate": 0.00016146190803019905, + "loss": 0.207, + "step": 37060 + }, + { + "epoch": 0.5782429649965682, + "grad_norm": 1.486069917678833, + "learning_rate": 0.00016145150891204427, + "loss": 0.1232, + "step": 37070 + }, + { + "epoch": 0.57839895176889, + "grad_norm": 0.13543163239955902, + "learning_rate": 0.0001614411097938895, + "loss": 0.3245, + "step": 37080 + }, + { + "epoch": 0.5785549385412118, + "grad_norm": 3.9703898429870605, + "learning_rate": 0.0001614307106757347, + "loss": 0.3104, + "step": 37090 + }, + { + "epoch": 0.5787109253135334, + "grad_norm": 3.2236313819885254, + "learning_rate": 0.00016142031155757993, + "loss": 0.1739, + "step": 37100 + }, + { + "epoch": 0.5788669120858552, + "grad_norm": 0.3854866325855255, + "learning_rate": 0.00016140991243942515, + "loss": 0.084, + "step": 37110 + }, + { + "epoch": 0.5790228988581768, + "grad_norm": 0.1981169581413269, + "learning_rate": 0.00016139951332127037, + "loss": 0.2367, + "step": 37120 + }, + { + "epoch": 0.5791788856304986, + "grad_norm": 2.7315785884857178, + "learning_rate": 0.0001613891142031156, + "loss": 0.1816, + "step": 37130 + }, + { + "epoch": 0.5793348724028202, + "grad_norm": 0.48251378536224365, + "learning_rate": 0.0001613787150849608, + "loss": 0.2123, + "step": 37140 + }, + { + "epoch": 0.579490859175142, + "grad_norm": 5.0431413650512695, + "learning_rate": 0.00016136831596680603, + "loss": 0.2489, + "step": 37150 + }, + { + "epoch": 0.5796468459474636, + "grad_norm": 0.81981360912323, + "learning_rate": 0.00016135791684865125, + "loss": 0.5235, + "step": 37160 + }, + { + "epoch": 0.5798028327197854, + "grad_norm": 0.5701965689659119, + "learning_rate": 0.00016134751773049647, + "loss": 0.1709, + "step": 37170 + }, + { + "epoch": 0.579958819492107, + "grad_norm": 1.3905616998672485, + "learning_rate": 0.00016133711861234169, + "loss": 0.2641, + "step": 37180 + }, + { + "epoch": 0.5801148062644288, + "grad_norm": 2.281285047531128, + "learning_rate": 0.0001613267194941869, + "loss": 0.3915, + "step": 37190 + }, + { + "epoch": 0.5802707930367504, + "grad_norm": 1.2038309574127197, + "learning_rate": 0.00016131632037603213, + "loss": 0.0881, + "step": 37200 + }, + { + "epoch": 0.5804267798090722, + "grad_norm": 0.3096400201320648, + "learning_rate": 0.00016130592125787735, + "loss": 0.1341, + "step": 37210 + }, + { + "epoch": 0.5805827665813938, + "grad_norm": 0.10219905525445938, + "learning_rate": 0.00016129552213972256, + "loss": 0.0867, + "step": 37220 + }, + { + "epoch": 0.5807387533537156, + "grad_norm": 0.13869646191596985, + "learning_rate": 0.00016128512302156778, + "loss": 0.1621, + "step": 37230 + }, + { + "epoch": 0.5808947401260373, + "grad_norm": 0.7992938160896301, + "learning_rate": 0.000161274723903413, + "loss": 0.275, + "step": 37240 + }, + { + "epoch": 0.581050726898359, + "grad_norm": 0.2635972201824188, + "learning_rate": 0.00016126432478525822, + "loss": 0.2429, + "step": 37250 + }, + { + "epoch": 0.5812067136706808, + "grad_norm": 1.7211238145828247, + "learning_rate": 0.00016125392566710344, + "loss": 0.1541, + "step": 37260 + }, + { + "epoch": 0.5813627004430024, + "grad_norm": 2.2448506355285645, + "learning_rate": 0.00016124352654894866, + "loss": 0.3689, + "step": 37270 + }, + { + "epoch": 0.5815186872153242, + "grad_norm": 1.238708257675171, + "learning_rate": 0.00016123312743079388, + "loss": 0.3292, + "step": 37280 + }, + { + "epoch": 0.5816746739876458, + "grad_norm": 0.5670503377914429, + "learning_rate": 0.0001612227283126391, + "loss": 0.1151, + "step": 37290 + }, + { + "epoch": 0.5818306607599676, + "grad_norm": 2.287304401397705, + "learning_rate": 0.00016121232919448432, + "loss": 0.3558, + "step": 37300 + }, + { + "epoch": 0.5819866475322892, + "grad_norm": 0.8009629845619202, + "learning_rate": 0.00016120193007632954, + "loss": 0.1861, + "step": 37310 + }, + { + "epoch": 0.582142634304611, + "grad_norm": 1.1901466846466064, + "learning_rate": 0.00016119153095817476, + "loss": 0.089, + "step": 37320 + }, + { + "epoch": 0.5822986210769326, + "grad_norm": 1.7198141813278198, + "learning_rate": 0.00016118113184001998, + "loss": 0.2931, + "step": 37330 + }, + { + "epoch": 0.5824546078492544, + "grad_norm": 0.8832749724388123, + "learning_rate": 0.0001611707327218652, + "loss": 0.2342, + "step": 37340 + }, + { + "epoch": 0.582610594621576, + "grad_norm": 1.0623290538787842, + "learning_rate": 0.00016116033360371042, + "loss": 0.2317, + "step": 37350 + }, + { + "epoch": 0.5827665813938978, + "grad_norm": 0.5173699855804443, + "learning_rate": 0.00016114993448555564, + "loss": 0.1299, + "step": 37360 + }, + { + "epoch": 0.5829225681662195, + "grad_norm": 0.05118720978498459, + "learning_rate": 0.00016113953536740086, + "loss": 0.0786, + "step": 37370 + }, + { + "epoch": 0.5830785549385412, + "grad_norm": 2.0060107707977295, + "learning_rate": 0.00016112913624924608, + "loss": 0.266, + "step": 37380 + }, + { + "epoch": 0.5832345417108629, + "grad_norm": 5.121693134307861, + "learning_rate": 0.0001611187371310913, + "loss": 0.1824, + "step": 37390 + }, + { + "epoch": 0.5833905284831846, + "grad_norm": 0.3568089008331299, + "learning_rate": 0.00016110833801293652, + "loss": 0.3189, + "step": 37400 + }, + { + "epoch": 0.5835465152555064, + "grad_norm": 0.8631492853164673, + "learning_rate": 0.00016109793889478174, + "loss": 0.1724, + "step": 37410 + }, + { + "epoch": 0.583702502027828, + "grad_norm": 1.9214404821395874, + "learning_rate": 0.00016108753977662693, + "loss": 0.2094, + "step": 37420 + }, + { + "epoch": 0.5838584888001498, + "grad_norm": 1.9055646657943726, + "learning_rate": 0.00016107714065847218, + "loss": 0.1258, + "step": 37430 + }, + { + "epoch": 0.5840144755724714, + "grad_norm": 0.6614194512367249, + "learning_rate": 0.00016106674154031737, + "loss": 0.1966, + "step": 37440 + }, + { + "epoch": 0.5841704623447932, + "grad_norm": 0.286883682012558, + "learning_rate": 0.00016105634242216262, + "loss": 0.588, + "step": 37450 + }, + { + "epoch": 0.5843264491171148, + "grad_norm": 0.5599405765533447, + "learning_rate": 0.0001610459433040078, + "loss": 0.3414, + "step": 37460 + }, + { + "epoch": 0.5844824358894366, + "grad_norm": 1.5908915996551514, + "learning_rate": 0.00016103554418585306, + "loss": 0.1264, + "step": 37470 + }, + { + "epoch": 0.5846384226617583, + "grad_norm": 1.0729930400848389, + "learning_rate": 0.00016102514506769825, + "loss": 0.4815, + "step": 37480 + }, + { + "epoch": 0.58479440943408, + "grad_norm": 0.8912356495857239, + "learning_rate": 0.0001610147459495435, + "loss": 0.2092, + "step": 37490 + }, + { + "epoch": 0.5849503962064017, + "grad_norm": 0.24116197228431702, + "learning_rate": 0.0001610043468313887, + "loss": 0.1552, + "step": 37500 + }, + { + "epoch": 0.5851063829787234, + "grad_norm": 2.245192527770996, + "learning_rate": 0.00016099394771323393, + "loss": 0.0553, + "step": 37510 + }, + { + "epoch": 0.5852623697510451, + "grad_norm": 0.11538795381784439, + "learning_rate": 0.00016098354859507913, + "loss": 0.0889, + "step": 37520 + }, + { + "epoch": 0.5854183565233668, + "grad_norm": 1.0372689962387085, + "learning_rate": 0.00016097314947692437, + "loss": 0.3909, + "step": 37530 + }, + { + "epoch": 0.5855743432956885, + "grad_norm": 1.1946550607681274, + "learning_rate": 0.00016096275035876957, + "loss": 0.2167, + "step": 37540 + }, + { + "epoch": 0.5857303300680102, + "grad_norm": 1.3474591970443726, + "learning_rate": 0.0001609523512406148, + "loss": 0.1561, + "step": 37550 + }, + { + "epoch": 0.585886316840332, + "grad_norm": 0.008936616592109203, + "learning_rate": 0.00016094195212246, + "loss": 0.3991, + "step": 37560 + }, + { + "epoch": 0.5860423036126536, + "grad_norm": 1.8822320699691772, + "learning_rate": 0.00016093155300430525, + "loss": 0.2438, + "step": 37570 + }, + { + "epoch": 0.5861982903849754, + "grad_norm": 1.1707524061203003, + "learning_rate": 0.00016092115388615044, + "loss": 0.308, + "step": 37580 + }, + { + "epoch": 0.586354277157297, + "grad_norm": 1.679641842842102, + "learning_rate": 0.0001609107547679957, + "loss": 0.159, + "step": 37590 + }, + { + "epoch": 0.5865102639296188, + "grad_norm": 1.6885284185409546, + "learning_rate": 0.00016090035564984088, + "loss": 0.157, + "step": 37600 + }, + { + "epoch": 0.5866662507019405, + "grad_norm": 0.3876996636390686, + "learning_rate": 0.00016088995653168613, + "loss": 0.4067, + "step": 37610 + }, + { + "epoch": 0.5868222374742622, + "grad_norm": 1.387480616569519, + "learning_rate": 0.00016087955741353132, + "loss": 0.1898, + "step": 37620 + }, + { + "epoch": 0.5869782242465839, + "grad_norm": 0.03653861582279205, + "learning_rate": 0.00016086915829537657, + "loss": 0.1652, + "step": 37630 + }, + { + "epoch": 0.5871342110189056, + "grad_norm": 0.6258412003517151, + "learning_rate": 0.00016085875917722176, + "loss": 0.2558, + "step": 37640 + }, + { + "epoch": 0.5872901977912273, + "grad_norm": 0.10519164055585861, + "learning_rate": 0.000160848360059067, + "loss": 0.243, + "step": 37650 + }, + { + "epoch": 0.587446184563549, + "grad_norm": 0.7370131015777588, + "learning_rate": 0.0001608379609409122, + "loss": 0.2343, + "step": 37660 + }, + { + "epoch": 0.5876021713358707, + "grad_norm": 0.6884517073631287, + "learning_rate": 0.00016082756182275745, + "loss": 0.1847, + "step": 37670 + }, + { + "epoch": 0.5877581581081924, + "grad_norm": 0.6326120495796204, + "learning_rate": 0.00016081716270460264, + "loss": 0.2258, + "step": 37680 + }, + { + "epoch": 0.5879141448805141, + "grad_norm": 7.697889804840088, + "learning_rate": 0.0001608067635864479, + "loss": 0.2068, + "step": 37690 + }, + { + "epoch": 0.5880701316528358, + "grad_norm": 1.8753420114517212, + "learning_rate": 0.00016079636446829308, + "loss": 0.3087, + "step": 37700 + }, + { + "epoch": 0.5882261184251576, + "grad_norm": 0.8591625094413757, + "learning_rate": 0.00016078596535013833, + "loss": 0.1082, + "step": 37710 + }, + { + "epoch": 0.5883821051974792, + "grad_norm": 0.6124423146247864, + "learning_rate": 0.00016077556623198352, + "loss": 0.1536, + "step": 37720 + }, + { + "epoch": 0.588538091969801, + "grad_norm": 0.1400165557861328, + "learning_rate": 0.00016076516711382877, + "loss": 0.145, + "step": 37730 + }, + { + "epoch": 0.5886940787421227, + "grad_norm": 1.9879004955291748, + "learning_rate": 0.00016075476799567396, + "loss": 0.4986, + "step": 37740 + }, + { + "epoch": 0.5888500655144444, + "grad_norm": 0.08687327802181244, + "learning_rate": 0.0001607443688775192, + "loss": 0.3365, + "step": 37750 + }, + { + "epoch": 0.5890060522867661, + "grad_norm": 0.3442278206348419, + "learning_rate": 0.0001607339697593644, + "loss": 0.2274, + "step": 37760 + }, + { + "epoch": 0.5891620390590878, + "grad_norm": 0.06124915927648544, + "learning_rate": 0.00016072357064120965, + "loss": 0.1915, + "step": 37770 + }, + { + "epoch": 0.5893180258314095, + "grad_norm": 1.0996816158294678, + "learning_rate": 0.00016071317152305484, + "loss": 0.1573, + "step": 37780 + }, + { + "epoch": 0.5894740126037312, + "grad_norm": 0.14738436043262482, + "learning_rate": 0.00016070277240490008, + "loss": 0.2524, + "step": 37790 + }, + { + "epoch": 0.5896299993760529, + "grad_norm": 2.3905467987060547, + "learning_rate": 0.00016069237328674528, + "loss": 0.1838, + "step": 37800 + }, + { + "epoch": 0.5897859861483746, + "grad_norm": 2.0800178050994873, + "learning_rate": 0.00016068197416859052, + "loss": 0.3749, + "step": 37810 + }, + { + "epoch": 0.5899419729206963, + "grad_norm": 3.254204511642456, + "learning_rate": 0.00016067157505043574, + "loss": 0.3777, + "step": 37820 + }, + { + "epoch": 0.590097959693018, + "grad_norm": 2.8447892665863037, + "learning_rate": 0.00016066117593228096, + "loss": 0.5338, + "step": 37830 + }, + { + "epoch": 0.5902539464653397, + "grad_norm": 0.624733030796051, + "learning_rate": 0.00016065077681412618, + "loss": 0.2331, + "step": 37840 + }, + { + "epoch": 0.5904099332376614, + "grad_norm": 0.6863507628440857, + "learning_rate": 0.0001606403776959714, + "loss": 0.1747, + "step": 37850 + }, + { + "epoch": 0.5905659200099832, + "grad_norm": 0.6785943508148193, + "learning_rate": 0.00016062997857781662, + "loss": 0.2064, + "step": 37860 + }, + { + "epoch": 0.5907219067823049, + "grad_norm": 0.5739938616752625, + "learning_rate": 0.00016061957945966181, + "loss": 0.2368, + "step": 37870 + }, + { + "epoch": 0.5908778935546266, + "grad_norm": 0.0616387277841568, + "learning_rate": 0.00016060918034150706, + "loss": 0.2224, + "step": 37880 + }, + { + "epoch": 0.5910338803269483, + "grad_norm": 2.745957136154175, + "learning_rate": 0.00016059878122335225, + "loss": 0.3161, + "step": 37890 + }, + { + "epoch": 0.59118986709927, + "grad_norm": 2.52472186088562, + "learning_rate": 0.0001605883821051975, + "loss": 0.1713, + "step": 37900 + }, + { + "epoch": 0.5913458538715917, + "grad_norm": 1.5691877603530884, + "learning_rate": 0.0001605779829870427, + "loss": 0.1402, + "step": 37910 + }, + { + "epoch": 0.5915018406439134, + "grad_norm": 1.848261833190918, + "learning_rate": 0.00016056758386888794, + "loss": 0.2305, + "step": 37920 + }, + { + "epoch": 0.5916578274162351, + "grad_norm": 0.36622482538223267, + "learning_rate": 0.00016055718475073313, + "loss": 0.1834, + "step": 37930 + }, + { + "epoch": 0.5918138141885568, + "grad_norm": 0.21201446652412415, + "learning_rate": 0.00016054678563257838, + "loss": 0.4682, + "step": 37940 + }, + { + "epoch": 0.5919698009608785, + "grad_norm": 0.38921502232551575, + "learning_rate": 0.00016053638651442357, + "loss": 0.2892, + "step": 37950 + }, + { + "epoch": 0.5921257877332002, + "grad_norm": 0.20777840912342072, + "learning_rate": 0.00016052598739626882, + "loss": 0.104, + "step": 37960 + }, + { + "epoch": 0.5922817745055219, + "grad_norm": 0.7090150713920593, + "learning_rate": 0.000160515588278114, + "loss": 0.2592, + "step": 37970 + }, + { + "epoch": 0.5924377612778436, + "grad_norm": 1.9939709901809692, + "learning_rate": 0.00016050518915995926, + "loss": 0.247, + "step": 37980 + }, + { + "epoch": 0.5925937480501653, + "grad_norm": 0.8696603178977966, + "learning_rate": 0.00016049479004180445, + "loss": 0.301, + "step": 37990 + }, + { + "epoch": 0.592749734822487, + "grad_norm": 0.4119095206260681, + "learning_rate": 0.0001604843909236497, + "loss": 0.2011, + "step": 38000 + }, + { + "epoch": 0.5929057215948088, + "grad_norm": 0.060975607484579086, + "learning_rate": 0.0001604739918054949, + "loss": 0.2286, + "step": 38010 + }, + { + "epoch": 0.5930617083671305, + "grad_norm": 0.7039875388145447, + "learning_rate": 0.00016046359268734014, + "loss": 0.0999, + "step": 38020 + }, + { + "epoch": 0.5932176951394522, + "grad_norm": 0.28402045369148254, + "learning_rate": 0.00016045319356918533, + "loss": 0.1835, + "step": 38030 + }, + { + "epoch": 0.5933736819117739, + "grad_norm": 2.217988967895508, + "learning_rate": 0.00016044279445103058, + "loss": 0.0785, + "step": 38040 + }, + { + "epoch": 0.5935296686840956, + "grad_norm": 1.4856069087982178, + "learning_rate": 0.00016043239533287577, + "loss": 0.2528, + "step": 38050 + }, + { + "epoch": 0.5936856554564173, + "grad_norm": 0.9875519871711731, + "learning_rate": 0.00016042199621472101, + "loss": 0.4445, + "step": 38060 + }, + { + "epoch": 0.593841642228739, + "grad_norm": 2.773216724395752, + "learning_rate": 0.0001604115970965662, + "loss": 0.5643, + "step": 38070 + }, + { + "epoch": 0.5939976290010607, + "grad_norm": 1.7584549188613892, + "learning_rate": 0.00016040119797841145, + "loss": 0.1446, + "step": 38080 + }, + { + "epoch": 0.5941536157733824, + "grad_norm": 1.2456908226013184, + "learning_rate": 0.00016039079886025665, + "loss": 0.1896, + "step": 38090 + }, + { + "epoch": 0.5943096025457041, + "grad_norm": 0.021098516881465912, + "learning_rate": 0.0001603803997421019, + "loss": 0.0876, + "step": 38100 + }, + { + "epoch": 0.5944655893180258, + "grad_norm": 1.8131749629974365, + "learning_rate": 0.00016037000062394709, + "loss": 0.1639, + "step": 38110 + }, + { + "epoch": 0.5946215760903475, + "grad_norm": 0.41819462180137634, + "learning_rate": 0.00016035960150579233, + "loss": 0.1311, + "step": 38120 + }, + { + "epoch": 0.5947775628626693, + "grad_norm": 0.6691219806671143, + "learning_rate": 0.00016034920238763753, + "loss": 0.1625, + "step": 38130 + }, + { + "epoch": 0.5949335496349909, + "grad_norm": 0.07854852080345154, + "learning_rate": 0.00016033880326948277, + "loss": 0.3177, + "step": 38140 + }, + { + "epoch": 0.5950895364073127, + "grad_norm": 0.7819689512252808, + "learning_rate": 0.00016032840415132796, + "loss": 0.4225, + "step": 38150 + }, + { + "epoch": 0.5952455231796344, + "grad_norm": 5.3962836265563965, + "learning_rate": 0.0001603180050331732, + "loss": 0.123, + "step": 38160 + }, + { + "epoch": 0.5954015099519561, + "grad_norm": 0.7460303902626038, + "learning_rate": 0.0001603076059150184, + "loss": 0.2182, + "step": 38170 + }, + { + "epoch": 0.5955574967242778, + "grad_norm": 0.26433685421943665, + "learning_rate": 0.00016029720679686365, + "loss": 0.1894, + "step": 38180 + }, + { + "epoch": 0.5957134834965995, + "grad_norm": 0.8701320290565491, + "learning_rate": 0.00016028680767870884, + "loss": 0.2137, + "step": 38190 + }, + { + "epoch": 0.5958694702689212, + "grad_norm": 3.9774930477142334, + "learning_rate": 0.0001602764085605541, + "loss": 0.4637, + "step": 38200 + }, + { + "epoch": 0.5960254570412429, + "grad_norm": 0.5801145434379578, + "learning_rate": 0.00016026600944239928, + "loss": 0.3036, + "step": 38210 + }, + { + "epoch": 0.5961814438135646, + "grad_norm": 1.1890723705291748, + "learning_rate": 0.00016025561032424453, + "loss": 0.0988, + "step": 38220 + }, + { + "epoch": 0.5963374305858863, + "grad_norm": 0.18521073460578918, + "learning_rate": 0.00016024521120608972, + "loss": 0.1529, + "step": 38230 + }, + { + "epoch": 0.596493417358208, + "grad_norm": 2.2602827548980713, + "learning_rate": 0.00016023481208793497, + "loss": 0.2815, + "step": 38240 + }, + { + "epoch": 0.5966494041305297, + "grad_norm": 0.4060989022254944, + "learning_rate": 0.00016022441296978016, + "loss": 0.3466, + "step": 38250 + }, + { + "epoch": 0.5968053909028515, + "grad_norm": 1.4743244647979736, + "learning_rate": 0.0001602140138516254, + "loss": 0.3561, + "step": 38260 + }, + { + "epoch": 0.5969613776751731, + "grad_norm": 5.080365180969238, + "learning_rate": 0.0001602036147334706, + "loss": 0.2914, + "step": 38270 + }, + { + "epoch": 0.5971173644474949, + "grad_norm": 0.16149812936782837, + "learning_rate": 0.00016019321561531585, + "loss": 0.1871, + "step": 38280 + }, + { + "epoch": 0.5972733512198165, + "grad_norm": 1.7535364627838135, + "learning_rate": 0.00016018281649716104, + "loss": 0.4905, + "step": 38290 + }, + { + "epoch": 0.5974293379921383, + "grad_norm": 0.4103987216949463, + "learning_rate": 0.00016017241737900629, + "loss": 0.3813, + "step": 38300 + }, + { + "epoch": 0.59758532476446, + "grad_norm": 0.5315369367599487, + "learning_rate": 0.00016016201826085148, + "loss": 0.3897, + "step": 38310 + }, + { + "epoch": 0.5977413115367817, + "grad_norm": 3.4553463459014893, + "learning_rate": 0.0001601516191426967, + "loss": 0.2047, + "step": 38320 + }, + { + "epoch": 0.5978972983091034, + "grad_norm": 2.225468397140503, + "learning_rate": 0.00016014122002454192, + "loss": 0.4028, + "step": 38330 + }, + { + "epoch": 0.5980532850814251, + "grad_norm": 2.000885486602783, + "learning_rate": 0.00016013082090638714, + "loss": 0.391, + "step": 38340 + }, + { + "epoch": 0.5982092718537468, + "grad_norm": 0.7886672616004944, + "learning_rate": 0.00016012042178823236, + "loss": 0.2994, + "step": 38350 + }, + { + "epoch": 0.5983652586260685, + "grad_norm": 0.5279917120933533, + "learning_rate": 0.00016011002267007758, + "loss": 0.1855, + "step": 38360 + }, + { + "epoch": 0.5985212453983902, + "grad_norm": 0.9789942502975464, + "learning_rate": 0.0001600996235519228, + "loss": 0.2122, + "step": 38370 + }, + { + "epoch": 0.5986772321707119, + "grad_norm": 0.6289750933647156, + "learning_rate": 0.00016008922443376802, + "loss": 0.3737, + "step": 38380 + }, + { + "epoch": 0.5988332189430337, + "grad_norm": 1.3074733018875122, + "learning_rate": 0.00016007882531561324, + "loss": 0.2382, + "step": 38390 + }, + { + "epoch": 0.5989892057153553, + "grad_norm": 1.745284914970398, + "learning_rate": 0.00016006842619745846, + "loss": 0.3167, + "step": 38400 + }, + { + "epoch": 0.5991451924876771, + "grad_norm": 0.1694452464580536, + "learning_rate": 0.00016005802707930368, + "loss": 0.1632, + "step": 38410 + }, + { + "epoch": 0.5993011792599987, + "grad_norm": 1.55560302734375, + "learning_rate": 0.0001600476279611489, + "loss": 0.2371, + "step": 38420 + }, + { + "epoch": 0.5994571660323205, + "grad_norm": 1.044997215270996, + "learning_rate": 0.00016003722884299411, + "loss": 0.0896, + "step": 38430 + }, + { + "epoch": 0.5996131528046421, + "grad_norm": 0.25250110030174255, + "learning_rate": 0.00016002682972483933, + "loss": 0.3096, + "step": 38440 + }, + { + "epoch": 0.5997691395769639, + "grad_norm": 0.1631859689950943, + "learning_rate": 0.00016001643060668455, + "loss": 0.2004, + "step": 38450 + }, + { + "epoch": 0.5999251263492856, + "grad_norm": 0.34967318177223206, + "learning_rate": 0.00016000603148852977, + "loss": 0.1476, + "step": 38460 + }, + { + "epoch": 0.6000811131216073, + "grad_norm": 1.8604000806808472, + "learning_rate": 0.000159995632370375, + "loss": 0.1884, + "step": 38470 + }, + { + "epoch": 0.600237099893929, + "grad_norm": 1.626848578453064, + "learning_rate": 0.0001599852332522202, + "loss": 0.5452, + "step": 38480 + }, + { + "epoch": 0.6003930866662507, + "grad_norm": 0.41935989260673523, + "learning_rate": 0.00015997483413406543, + "loss": 0.3047, + "step": 38490 + }, + { + "epoch": 0.6005490734385724, + "grad_norm": 1.355086088180542, + "learning_rate": 0.00015996443501591065, + "loss": 0.1246, + "step": 38500 + }, + { + "epoch": 0.6007050602108941, + "grad_norm": 1.5382829904556274, + "learning_rate": 0.00015995403589775587, + "loss": 0.2324, + "step": 38510 + }, + { + "epoch": 0.6008610469832159, + "grad_norm": 1.0250797271728516, + "learning_rate": 0.0001599436367796011, + "loss": 0.181, + "step": 38520 + }, + { + "epoch": 0.6010170337555375, + "grad_norm": 0.07823364436626434, + "learning_rate": 0.0001599332376614463, + "loss": 0.1492, + "step": 38530 + }, + { + "epoch": 0.6011730205278593, + "grad_norm": 0.07573071867227554, + "learning_rate": 0.00015992283854329153, + "loss": 0.1256, + "step": 38540 + }, + { + "epoch": 0.6013290073001809, + "grad_norm": 0.5608569383621216, + "learning_rate": 0.00015991243942513675, + "loss": 0.378, + "step": 38550 + }, + { + "epoch": 0.6014849940725027, + "grad_norm": 0.7001075744628906, + "learning_rate": 0.00015990204030698197, + "loss": 0.2518, + "step": 38560 + }, + { + "epoch": 0.6016409808448243, + "grad_norm": 4.868145942687988, + "learning_rate": 0.0001598916411888272, + "loss": 0.1857, + "step": 38570 + }, + { + "epoch": 0.6017969676171461, + "grad_norm": 0.05653705075383186, + "learning_rate": 0.0001598812420706724, + "loss": 0.1918, + "step": 38580 + }, + { + "epoch": 0.6019529543894677, + "grad_norm": 0.8556832671165466, + "learning_rate": 0.00015987084295251763, + "loss": 0.1926, + "step": 38590 + }, + { + "epoch": 0.6021089411617895, + "grad_norm": 1.1988335847854614, + "learning_rate": 0.00015986044383436285, + "loss": 0.3884, + "step": 38600 + }, + { + "epoch": 0.6022649279341112, + "grad_norm": 1.5978832244873047, + "learning_rate": 0.00015985004471620807, + "loss": 0.1986, + "step": 38610 + }, + { + "epoch": 0.6024209147064329, + "grad_norm": 1.084174633026123, + "learning_rate": 0.0001598396455980533, + "loss": 0.4602, + "step": 38620 + }, + { + "epoch": 0.6025769014787546, + "grad_norm": 1.5116629600524902, + "learning_rate": 0.0001598292464798985, + "loss": 0.292, + "step": 38630 + }, + { + "epoch": 0.6027328882510763, + "grad_norm": 0.23945322632789612, + "learning_rate": 0.00015981884736174373, + "loss": 0.1904, + "step": 38640 + }, + { + "epoch": 0.602888875023398, + "grad_norm": 0.348156601190567, + "learning_rate": 0.00015980844824358895, + "loss": 0.1442, + "step": 38650 + }, + { + "epoch": 0.6030448617957197, + "grad_norm": 1.876736044883728, + "learning_rate": 0.00015979804912543417, + "loss": 0.248, + "step": 38660 + }, + { + "epoch": 0.6032008485680415, + "grad_norm": 0.2564373016357422, + "learning_rate": 0.00015978765000727939, + "loss": 0.2429, + "step": 38670 + }, + { + "epoch": 0.6033568353403631, + "grad_norm": 1.824837327003479, + "learning_rate": 0.0001597772508891246, + "loss": 0.2593, + "step": 38680 + }, + { + "epoch": 0.6035128221126849, + "grad_norm": 1.6967028379440308, + "learning_rate": 0.00015976685177096985, + "loss": 0.3918, + "step": 38690 + }, + { + "epoch": 0.6036688088850065, + "grad_norm": 1.6629718542099, + "learning_rate": 0.00015975645265281504, + "loss": 0.2529, + "step": 38700 + }, + { + "epoch": 0.6038247956573283, + "grad_norm": 0.19047370553016663, + "learning_rate": 0.0001597460535346603, + "loss": 0.4545, + "step": 38710 + }, + { + "epoch": 0.6039807824296499, + "grad_norm": 0.6910511255264282, + "learning_rate": 0.00015973565441650548, + "loss": 0.2226, + "step": 38720 + }, + { + "epoch": 0.6041367692019717, + "grad_norm": 1.9387328624725342, + "learning_rate": 0.00015972525529835073, + "loss": 0.5561, + "step": 38730 + }, + { + "epoch": 0.6042927559742933, + "grad_norm": 1.4491708278656006, + "learning_rate": 0.00015971485618019592, + "loss": 0.2987, + "step": 38740 + }, + { + "epoch": 0.6044487427466151, + "grad_norm": 0.32534459233283997, + "learning_rate": 0.00015970445706204117, + "loss": 0.1683, + "step": 38750 + }, + { + "epoch": 0.6046047295189368, + "grad_norm": 0.17335011065006256, + "learning_rate": 0.00015969405794388636, + "loss": 0.1554, + "step": 38760 + }, + { + "epoch": 0.6047607162912585, + "grad_norm": 0.39481663703918457, + "learning_rate": 0.00015968365882573158, + "loss": 0.186, + "step": 38770 + }, + { + "epoch": 0.6049167030635803, + "grad_norm": 0.10422719269990921, + "learning_rate": 0.0001596732597075768, + "loss": 0.1599, + "step": 38780 + }, + { + "epoch": 0.6050726898359019, + "grad_norm": 2.288064479827881, + "learning_rate": 0.00015966286058942202, + "loss": 0.1944, + "step": 38790 + }, + { + "epoch": 0.6052286766082237, + "grad_norm": 1.0062401294708252, + "learning_rate": 0.00015965246147126724, + "loss": 0.176, + "step": 38800 + }, + { + "epoch": 0.6053846633805453, + "grad_norm": 0.21470843255519867, + "learning_rate": 0.00015964206235311246, + "loss": 0.183, + "step": 38810 + }, + { + "epoch": 0.6055406501528671, + "grad_norm": 2.1730270385742188, + "learning_rate": 0.00015963166323495768, + "loss": 0.1788, + "step": 38820 + }, + { + "epoch": 0.6056966369251887, + "grad_norm": 0.6188749074935913, + "learning_rate": 0.0001596212641168029, + "loss": 0.2746, + "step": 38830 + }, + { + "epoch": 0.6058526236975105, + "grad_norm": 1.283058524131775, + "learning_rate": 0.00015961086499864812, + "loss": 0.1066, + "step": 38840 + }, + { + "epoch": 0.6060086104698321, + "grad_norm": 1.1469789743423462, + "learning_rate": 0.00015960046588049334, + "loss": 0.0692, + "step": 38850 + }, + { + "epoch": 0.6061645972421539, + "grad_norm": 1.9450418949127197, + "learning_rate": 0.00015959006676233856, + "loss": 0.2154, + "step": 38860 + }, + { + "epoch": 0.6063205840144755, + "grad_norm": 7.193401336669922, + "learning_rate": 0.00015957966764418378, + "loss": 0.218, + "step": 38870 + }, + { + "epoch": 0.6064765707867973, + "grad_norm": 3.5315003395080566, + "learning_rate": 0.000159569268526029, + "loss": 0.312, + "step": 38880 + }, + { + "epoch": 0.6066325575591189, + "grad_norm": 1.2873650789260864, + "learning_rate": 0.00015955886940787422, + "loss": 0.2587, + "step": 38890 + }, + { + "epoch": 0.6067885443314407, + "grad_norm": 1.1193941831588745, + "learning_rate": 0.00015954847028971944, + "loss": 0.2358, + "step": 38900 + }, + { + "epoch": 0.6069445311037625, + "grad_norm": 1.0471165180206299, + "learning_rate": 0.00015953807117156466, + "loss": 0.2639, + "step": 38910 + }, + { + "epoch": 0.6071005178760841, + "grad_norm": 1.7760034799575806, + "learning_rate": 0.00015952767205340988, + "loss": 0.248, + "step": 38920 + }, + { + "epoch": 0.6072565046484059, + "grad_norm": 0.050742994993925095, + "learning_rate": 0.0001595172729352551, + "loss": 0.1287, + "step": 38930 + }, + { + "epoch": 0.6074124914207275, + "grad_norm": 0.03016967698931694, + "learning_rate": 0.00015950687381710032, + "loss": 0.1843, + "step": 38940 + }, + { + "epoch": 0.6075684781930493, + "grad_norm": 0.6029847860336304, + "learning_rate": 0.00015949647469894554, + "loss": 0.2294, + "step": 38950 + }, + { + "epoch": 0.6077244649653709, + "grad_norm": 0.2955610454082489, + "learning_rate": 0.00015948607558079076, + "loss": 0.3336, + "step": 38960 + }, + { + "epoch": 0.6078804517376927, + "grad_norm": 1.9104161262512207, + "learning_rate": 0.00015947567646263597, + "loss": 0.189, + "step": 38970 + }, + { + "epoch": 0.6080364385100143, + "grad_norm": 0.09197711199522018, + "learning_rate": 0.0001594652773444812, + "loss": 0.1612, + "step": 38980 + }, + { + "epoch": 0.6081924252823361, + "grad_norm": 1.4233628511428833, + "learning_rate": 0.00015945487822632641, + "loss": 0.1521, + "step": 38990 + }, + { + "epoch": 0.6083484120546577, + "grad_norm": 9.787493705749512, + "learning_rate": 0.00015944447910817163, + "loss": 0.135, + "step": 39000 + }, + { + "epoch": 0.6085043988269795, + "grad_norm": 1.3848742246627808, + "learning_rate": 0.00015943407999001685, + "loss": 0.2268, + "step": 39010 + }, + { + "epoch": 0.6086603855993011, + "grad_norm": 0.1524626612663269, + "learning_rate": 0.00015942368087186207, + "loss": 0.3154, + "step": 39020 + }, + { + "epoch": 0.6088163723716229, + "grad_norm": 2.1885435581207275, + "learning_rate": 0.0001594132817537073, + "loss": 0.3174, + "step": 39030 + }, + { + "epoch": 0.6089723591439445, + "grad_norm": 1.4719898700714111, + "learning_rate": 0.0001594028826355525, + "loss": 0.2959, + "step": 39040 + }, + { + "epoch": 0.6091283459162663, + "grad_norm": 0.7844040989875793, + "learning_rate": 0.00015939248351739773, + "loss": 0.3021, + "step": 39050 + }, + { + "epoch": 0.6092843326885881, + "grad_norm": 0.17660890519618988, + "learning_rate": 0.00015938208439924295, + "loss": 0.3608, + "step": 39060 + }, + { + "epoch": 0.6094403194609097, + "grad_norm": 0.24621741473674774, + "learning_rate": 0.00015937168528108817, + "loss": 0.1461, + "step": 39070 + }, + { + "epoch": 0.6095963062332315, + "grad_norm": 0.7903050184249878, + "learning_rate": 0.0001593612861629334, + "loss": 0.1195, + "step": 39080 + }, + { + "epoch": 0.6097522930055531, + "grad_norm": 0.04043950140476227, + "learning_rate": 0.0001593508870447786, + "loss": 0.1938, + "step": 39090 + }, + { + "epoch": 0.6099082797778749, + "grad_norm": 0.7602111101150513, + "learning_rate": 0.00015934048792662383, + "loss": 0.1732, + "step": 39100 + }, + { + "epoch": 0.6100642665501965, + "grad_norm": 0.7902756333351135, + "learning_rate": 0.00015933008880846905, + "loss": 0.2351, + "step": 39110 + }, + { + "epoch": 0.6102202533225183, + "grad_norm": 1.7035608291625977, + "learning_rate": 0.00015931968969031427, + "loss": 0.2409, + "step": 39120 + }, + { + "epoch": 0.6103762400948399, + "grad_norm": 0.7998217940330505, + "learning_rate": 0.0001593092905721595, + "loss": 0.3862, + "step": 39130 + }, + { + "epoch": 0.6105322268671617, + "grad_norm": 0.3530103862285614, + "learning_rate": 0.0001592988914540047, + "loss": 0.2176, + "step": 39140 + }, + { + "epoch": 0.6106882136394833, + "grad_norm": 0.9218409657478333, + "learning_rate": 0.00015928849233584993, + "loss": 0.1362, + "step": 39150 + }, + { + "epoch": 0.6108442004118051, + "grad_norm": 0.643829345703125, + "learning_rate": 0.00015927809321769515, + "loss": 0.2309, + "step": 39160 + }, + { + "epoch": 0.6110001871841267, + "grad_norm": 0.1810176521539688, + "learning_rate": 0.00015926769409954037, + "loss": 0.2461, + "step": 39170 + }, + { + "epoch": 0.6111561739564485, + "grad_norm": 1.3435649871826172, + "learning_rate": 0.0001592572949813856, + "loss": 0.2937, + "step": 39180 + }, + { + "epoch": 0.6113121607287701, + "grad_norm": 1.1048580408096313, + "learning_rate": 0.0001592468958632308, + "loss": 0.2983, + "step": 39190 + }, + { + "epoch": 0.6114681475010919, + "grad_norm": 0.09463644027709961, + "learning_rate": 0.00015923649674507603, + "loss": 0.2318, + "step": 39200 + }, + { + "epoch": 0.6116241342734137, + "grad_norm": 0.9329594969749451, + "learning_rate": 0.00015922609762692125, + "loss": 0.3095, + "step": 39210 + }, + { + "epoch": 0.6117801210457353, + "grad_norm": 1.8576663732528687, + "learning_rate": 0.00015921569850876647, + "loss": 0.2336, + "step": 39220 + }, + { + "epoch": 0.6119361078180571, + "grad_norm": 1.6159908771514893, + "learning_rate": 0.00015920529939061169, + "loss": 0.1745, + "step": 39230 + }, + { + "epoch": 0.6120920945903787, + "grad_norm": 1.289712905883789, + "learning_rate": 0.0001591949002724569, + "loss": 0.2031, + "step": 39240 + }, + { + "epoch": 0.6122480813627005, + "grad_norm": 0.4183761179447174, + "learning_rate": 0.00015918450115430212, + "loss": 0.2911, + "step": 39250 + }, + { + "epoch": 0.6124040681350221, + "grad_norm": 2.125042200088501, + "learning_rate": 0.00015917410203614734, + "loss": 0.1086, + "step": 39260 + }, + { + "epoch": 0.6125600549073439, + "grad_norm": 3.527683973312378, + "learning_rate": 0.00015916370291799256, + "loss": 0.3356, + "step": 39270 + }, + { + "epoch": 0.6127160416796655, + "grad_norm": 1.1525741815567017, + "learning_rate": 0.00015915330379983778, + "loss": 0.222, + "step": 39280 + }, + { + "epoch": 0.6128720284519873, + "grad_norm": 0.5210650563240051, + "learning_rate": 0.000159142904681683, + "loss": 0.1563, + "step": 39290 + }, + { + "epoch": 0.6130280152243089, + "grad_norm": 0.2881133258342743, + "learning_rate": 0.00015913250556352822, + "loss": 0.346, + "step": 39300 + }, + { + "epoch": 0.6131840019966307, + "grad_norm": 1.7034498453140259, + "learning_rate": 0.00015912210644537344, + "loss": 0.1822, + "step": 39310 + }, + { + "epoch": 0.6133399887689523, + "grad_norm": 2.6305196285247803, + "learning_rate": 0.00015911170732721866, + "loss": 0.3373, + "step": 39320 + }, + { + "epoch": 0.6134959755412741, + "grad_norm": 0.4541589915752411, + "learning_rate": 0.00015910130820906388, + "loss": 0.2146, + "step": 39330 + }, + { + "epoch": 0.6136519623135958, + "grad_norm": 0.03301222622394562, + "learning_rate": 0.0001590909090909091, + "loss": 0.1787, + "step": 39340 + }, + { + "epoch": 0.6138079490859175, + "grad_norm": 0.7106437087059021, + "learning_rate": 0.00015908050997275432, + "loss": 0.2315, + "step": 39350 + }, + { + "epoch": 0.6139639358582393, + "grad_norm": 0.381320059299469, + "learning_rate": 0.00015907011085459954, + "loss": 0.1917, + "step": 39360 + }, + { + "epoch": 0.6141199226305609, + "grad_norm": 0.5418739914894104, + "learning_rate": 0.00015905971173644476, + "loss": 0.1571, + "step": 39370 + }, + { + "epoch": 0.6142759094028827, + "grad_norm": 0.3135250210762024, + "learning_rate": 0.00015904931261828998, + "loss": 0.0858, + "step": 39380 + }, + { + "epoch": 0.6144318961752043, + "grad_norm": 1.6296218633651733, + "learning_rate": 0.0001590389135001352, + "loss": 0.2297, + "step": 39390 + }, + { + "epoch": 0.6145878829475261, + "grad_norm": 0.5768391489982605, + "learning_rate": 0.00015902851438198042, + "loss": 0.1699, + "step": 39400 + }, + { + "epoch": 0.6147438697198477, + "grad_norm": 25.794370651245117, + "learning_rate": 0.00015901811526382564, + "loss": 0.1903, + "step": 39410 + }, + { + "epoch": 0.6148998564921695, + "grad_norm": 0.46698322892189026, + "learning_rate": 0.00015900771614567086, + "loss": 0.1415, + "step": 39420 + }, + { + "epoch": 0.6150558432644911, + "grad_norm": 1.3042336702346802, + "learning_rate": 0.00015899731702751608, + "loss": 0.1604, + "step": 39430 + }, + { + "epoch": 0.6152118300368129, + "grad_norm": 1.2311556339263916, + "learning_rate": 0.0001589869179093613, + "loss": 0.2555, + "step": 39440 + }, + { + "epoch": 0.6153678168091345, + "grad_norm": 0.6290935277938843, + "learning_rate": 0.00015897651879120652, + "loss": 0.1723, + "step": 39450 + }, + { + "epoch": 0.6155238035814563, + "grad_norm": 1.4786419868469238, + "learning_rate": 0.00015896611967305174, + "loss": 0.3717, + "step": 39460 + }, + { + "epoch": 0.615679790353778, + "grad_norm": 0.7589457631111145, + "learning_rate": 0.00015895572055489696, + "loss": 0.1912, + "step": 39470 + }, + { + "epoch": 0.6158357771260997, + "grad_norm": 0.10508652776479721, + "learning_rate": 0.00015894532143674218, + "loss": 0.1353, + "step": 39480 + }, + { + "epoch": 0.6159917638984214, + "grad_norm": 0.07394446432590485, + "learning_rate": 0.0001589349223185874, + "loss": 0.1195, + "step": 39490 + }, + { + "epoch": 0.6161477506707431, + "grad_norm": 0.11420007050037384, + "learning_rate": 0.00015892452320043262, + "loss": 0.0976, + "step": 39500 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.405383667712e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-39500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2517f0d95e6cc6841351be1c3ef39fb39f1ab96a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a766f94894bb6da757f0c5e48ae017c732dacca361599d8561d282286c8f616 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..f1e4a63d17af8ddbe9445f4867d5d5cf42d6f875 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:74a04b7151d5a10ed219c55cb6f91ae07c7c7ffbc08149da765fbbdd51c2ab9f +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..d4cd97b786defba8a34f869dd00569a133cdd42a Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..caabd0af3859271c22e3c532c3b0c9aa8c519c2c Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..06fb8aa4e562aeec0d4fe3292d7453d0e61ac30e --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/trainer_state.json @@ -0,0 +1,28034 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.6239470892868285, + "eval_steps": 500, + "global_step": 40000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + }, + { + "epoch": 0.5071129968178698, + "grad_norm": 1.0075199604034424, + "learning_rate": 0.00016619350679062414, + "loss": 0.2569, + "step": 32510 + }, + { + "epoch": 0.5072689835901916, + "grad_norm": 1.3489066362380981, + "learning_rate": 0.0001661831076724694, + "loss": 0.2565, + "step": 32520 + }, + { + "epoch": 0.5074249703625132, + "grad_norm": 0.2534058094024658, + "learning_rate": 0.00016617270855431458, + "loss": 0.2292, + "step": 32530 + }, + { + "epoch": 0.507580957134835, + "grad_norm": 0.31723281741142273, + "learning_rate": 0.00016616230943615983, + "loss": 0.1467, + "step": 32540 + }, + { + "epoch": 0.5077369439071566, + "grad_norm": 0.29659849405288696, + "learning_rate": 0.00016615191031800502, + "loss": 0.2768, + "step": 32550 + }, + { + "epoch": 0.5078929306794784, + "grad_norm": 1.5246435403823853, + "learning_rate": 0.00016614151119985027, + "loss": 0.216, + "step": 32560 + }, + { + "epoch": 0.5080489174518001, + "grad_norm": 0.6026854515075684, + "learning_rate": 0.00016613111208169546, + "loss": 0.2277, + "step": 32570 + }, + { + "epoch": 0.5082049042241218, + "grad_norm": 0.41994890570640564, + "learning_rate": 0.0001661207129635407, + "loss": 0.2433, + "step": 32580 + }, + { + "epoch": 0.5083608909964435, + "grad_norm": 2.084599733352661, + "learning_rate": 0.00016611031384538593, + "loss": 0.4081, + "step": 32590 + }, + { + "epoch": 0.5085168777687652, + "grad_norm": 0.2945229709148407, + "learning_rate": 0.00016609991472723115, + "loss": 0.1329, + "step": 32600 + }, + { + "epoch": 0.508672864541087, + "grad_norm": 1.4981423616409302, + "learning_rate": 0.00016608951560907636, + "loss": 0.2855, + "step": 32610 + }, + { + "epoch": 0.5088288513134086, + "grad_norm": 0.961737871170044, + "learning_rate": 0.00016607911649092158, + "loss": 0.1654, + "step": 32620 + }, + { + "epoch": 0.5089848380857304, + "grad_norm": 1.49734628200531, + "learning_rate": 0.0001660687173727668, + "loss": 0.2081, + "step": 32630 + }, + { + "epoch": 0.509140824858052, + "grad_norm": 2.538135051727295, + "learning_rate": 0.00016605831825461202, + "loss": 0.3023, + "step": 32640 + }, + { + "epoch": 0.5092968116303738, + "grad_norm": 0.2981272041797638, + "learning_rate": 0.00016604791913645724, + "loss": 0.214, + "step": 32650 + }, + { + "epoch": 0.5094527984026954, + "grad_norm": 1.8863650560379028, + "learning_rate": 0.00016603752001830246, + "loss": 0.3363, + "step": 32660 + }, + { + "epoch": 0.5096087851750172, + "grad_norm": 0.022720983251929283, + "learning_rate": 0.00016602712090014768, + "loss": 0.2463, + "step": 32670 + }, + { + "epoch": 0.5097647719473388, + "grad_norm": 1.154977798461914, + "learning_rate": 0.0001660167217819929, + "loss": 0.4484, + "step": 32680 + }, + { + "epoch": 0.5099207587196606, + "grad_norm": 0.9470499157905579, + "learning_rate": 0.00016600632266383812, + "loss": 0.1196, + "step": 32690 + }, + { + "epoch": 0.5100767454919822, + "grad_norm": 0.19955401122570038, + "learning_rate": 0.00016599592354568334, + "loss": 0.1762, + "step": 32700 + }, + { + "epoch": 0.510232732264304, + "grad_norm": 0.4098282754421234, + "learning_rate": 0.00016598552442752856, + "loss": 0.2739, + "step": 32710 + }, + { + "epoch": 0.5103887190366257, + "grad_norm": 0.21697258949279785, + "learning_rate": 0.00016597512530937378, + "loss": 0.2555, + "step": 32720 + }, + { + "epoch": 0.5105447058089474, + "grad_norm": 4.894479751586914, + "learning_rate": 0.000165964726191219, + "loss": 0.3137, + "step": 32730 + }, + { + "epoch": 0.5107006925812692, + "grad_norm": 0.8396711349487305, + "learning_rate": 0.00016595432707306422, + "loss": 0.3105, + "step": 32740 + }, + { + "epoch": 0.5108566793535908, + "grad_norm": 1.5494674444198608, + "learning_rate": 0.00016594392795490944, + "loss": 0.2271, + "step": 32750 + }, + { + "epoch": 0.5110126661259126, + "grad_norm": 1.3409888744354248, + "learning_rate": 0.00016593352883675466, + "loss": 0.1796, + "step": 32760 + }, + { + "epoch": 0.5111686528982342, + "grad_norm": 1.7590911388397217, + "learning_rate": 0.00016592312971859988, + "loss": 0.2185, + "step": 32770 + }, + { + "epoch": 0.511324639670556, + "grad_norm": 0.7350002527236938, + "learning_rate": 0.0001659127306004451, + "loss": 0.1911, + "step": 32780 + }, + { + "epoch": 0.5114806264428776, + "grad_norm": 0.7176411747932434, + "learning_rate": 0.00016590233148229032, + "loss": 0.6918, + "step": 32790 + }, + { + "epoch": 0.5116366132151994, + "grad_norm": 2.5005321502685547, + "learning_rate": 0.00016589193236413554, + "loss": 0.2329, + "step": 32800 + }, + { + "epoch": 0.511792599987521, + "grad_norm": 1.3916250467300415, + "learning_rate": 0.00016588153324598076, + "loss": 0.2102, + "step": 32810 + }, + { + "epoch": 0.5119485867598428, + "grad_norm": 0.721671998500824, + "learning_rate": 0.00016587113412782598, + "loss": 0.2965, + "step": 32820 + }, + { + "epoch": 0.5121045735321644, + "grad_norm": 0.6120114326477051, + "learning_rate": 0.0001658607350096712, + "loss": 0.2661, + "step": 32830 + }, + { + "epoch": 0.5122605603044862, + "grad_norm": 0.29123347997665405, + "learning_rate": 0.00016585033589151642, + "loss": 0.2455, + "step": 32840 + }, + { + "epoch": 0.5124165470768078, + "grad_norm": 2.7300868034362793, + "learning_rate": 0.00016583993677336164, + "loss": 0.2049, + "step": 32850 + }, + { + "epoch": 0.5125725338491296, + "grad_norm": 1.801928997039795, + "learning_rate": 0.00016582953765520686, + "loss": 0.1725, + "step": 32860 + }, + { + "epoch": 0.5127285206214514, + "grad_norm": 0.6406471729278564, + "learning_rate": 0.00016581913853705208, + "loss": 0.164, + "step": 32870 + }, + { + "epoch": 0.512884507393773, + "grad_norm": 0.9627325534820557, + "learning_rate": 0.0001658087394188973, + "loss": 0.1728, + "step": 32880 + }, + { + "epoch": 0.5130404941660948, + "grad_norm": 1.5446962118148804, + "learning_rate": 0.00016579834030074251, + "loss": 0.2345, + "step": 32890 + }, + { + "epoch": 0.5131964809384164, + "grad_norm": 0.28092706203460693, + "learning_rate": 0.0001657879411825877, + "loss": 0.2124, + "step": 32900 + }, + { + "epoch": 0.5133524677107382, + "grad_norm": 1.7751421928405762, + "learning_rate": 0.00016577754206443295, + "loss": 0.2367, + "step": 32910 + }, + { + "epoch": 0.5135084544830598, + "grad_norm": 1.6964634656906128, + "learning_rate": 0.00016576714294627815, + "loss": 0.2361, + "step": 32920 + }, + { + "epoch": 0.5136644412553816, + "grad_norm": 0.06759823113679886, + "learning_rate": 0.0001657567438281234, + "loss": 0.2329, + "step": 32930 + }, + { + "epoch": 0.5138204280277032, + "grad_norm": 2.2519280910491943, + "learning_rate": 0.00016574634470996859, + "loss": 0.1079, + "step": 32940 + }, + { + "epoch": 0.513976414800025, + "grad_norm": 3.167576551437378, + "learning_rate": 0.00016573594559181383, + "loss": 0.1764, + "step": 32950 + }, + { + "epoch": 0.5141324015723466, + "grad_norm": 1.6734397411346436, + "learning_rate": 0.00016572554647365903, + "loss": 0.2878, + "step": 32960 + }, + { + "epoch": 0.5142883883446684, + "grad_norm": 0.35627374053001404, + "learning_rate": 0.00016571514735550427, + "loss": 0.2337, + "step": 32970 + }, + { + "epoch": 0.51444437511699, + "grad_norm": 1.6321039199829102, + "learning_rate": 0.00016570474823734946, + "loss": 0.2551, + "step": 32980 + }, + { + "epoch": 0.5146003618893118, + "grad_norm": 1.352982997894287, + "learning_rate": 0.0001656943491191947, + "loss": 0.4467, + "step": 32990 + }, + { + "epoch": 0.5147563486616334, + "grad_norm": 0.3558129072189331, + "learning_rate": 0.0001656839500010399, + "loss": 0.1913, + "step": 33000 + }, + { + "epoch": 0.5149123354339552, + "grad_norm": 0.041972286999225616, + "learning_rate": 0.00016567355088288515, + "loss": 0.1704, + "step": 33010 + }, + { + "epoch": 0.515068322206277, + "grad_norm": 1.4168939590454102, + "learning_rate": 0.00016566315176473034, + "loss": 0.4866, + "step": 33020 + }, + { + "epoch": 0.5152243089785986, + "grad_norm": 3.0042929649353027, + "learning_rate": 0.0001656527526465756, + "loss": 0.2473, + "step": 33030 + }, + { + "epoch": 0.5153802957509204, + "grad_norm": 1.027274489402771, + "learning_rate": 0.00016564235352842078, + "loss": 0.436, + "step": 33040 + }, + { + "epoch": 0.515536282523242, + "grad_norm": 1.2330195903778076, + "learning_rate": 0.00016563195441026603, + "loss": 0.4615, + "step": 33050 + }, + { + "epoch": 0.5156922692955638, + "grad_norm": 2.7145066261291504, + "learning_rate": 0.00016562155529211122, + "loss": 0.278, + "step": 33060 + }, + { + "epoch": 0.5158482560678854, + "grad_norm": 1.1273893117904663, + "learning_rate": 0.00016561115617395647, + "loss": 0.3767, + "step": 33070 + }, + { + "epoch": 0.5160042428402072, + "grad_norm": 2.466265916824341, + "learning_rate": 0.00016560075705580166, + "loss": 0.1879, + "step": 33080 + }, + { + "epoch": 0.5161602296125288, + "grad_norm": 0.3759515583515167, + "learning_rate": 0.0001655903579376469, + "loss": 0.4004, + "step": 33090 + }, + { + "epoch": 0.5163162163848506, + "grad_norm": 1.8821336030960083, + "learning_rate": 0.0001655799588194921, + "loss": 0.1622, + "step": 33100 + }, + { + "epoch": 0.5164722031571722, + "grad_norm": 0.8162096738815308, + "learning_rate": 0.00016556955970133735, + "loss": 0.1341, + "step": 33110 + }, + { + "epoch": 0.516628189929494, + "grad_norm": 0.008498801849782467, + "learning_rate": 0.00016555916058318254, + "loss": 0.2546, + "step": 33120 + }, + { + "epoch": 0.5167841767018156, + "grad_norm": 0.17536675930023193, + "learning_rate": 0.00016554876146502779, + "loss": 0.2934, + "step": 33130 + }, + { + "epoch": 0.5169401634741374, + "grad_norm": 3.539898633956909, + "learning_rate": 0.00016553836234687298, + "loss": 0.4959, + "step": 33140 + }, + { + "epoch": 0.517096150246459, + "grad_norm": 2.190887212753296, + "learning_rate": 0.00016552796322871823, + "loss": 0.2914, + "step": 33150 + }, + { + "epoch": 0.5172521370187808, + "grad_norm": 1.6388943195343018, + "learning_rate": 0.00016551756411056342, + "loss": 0.1317, + "step": 33160 + }, + { + "epoch": 0.5174081237911026, + "grad_norm": 0.33063387870788574, + "learning_rate": 0.00016550716499240866, + "loss": 0.4258, + "step": 33170 + }, + { + "epoch": 0.5175641105634242, + "grad_norm": 0.5168288946151733, + "learning_rate": 0.00016549676587425386, + "loss": 0.2548, + "step": 33180 + }, + { + "epoch": 0.517720097335746, + "grad_norm": 3.9570281505584717, + "learning_rate": 0.0001654863667560991, + "loss": 0.3179, + "step": 33190 + }, + { + "epoch": 0.5178760841080676, + "grad_norm": 1.127255916595459, + "learning_rate": 0.0001654759676379443, + "loss": 0.233, + "step": 33200 + }, + { + "epoch": 0.5180320708803894, + "grad_norm": 0.36299410462379456, + "learning_rate": 0.00016546556851978954, + "loss": 0.4614, + "step": 33210 + }, + { + "epoch": 0.518188057652711, + "grad_norm": 2.914668321609497, + "learning_rate": 0.00016545516940163474, + "loss": 0.1603, + "step": 33220 + }, + { + "epoch": 0.5183440444250328, + "grad_norm": 6.991094589233398, + "learning_rate": 0.00016544477028347998, + "loss": 0.2259, + "step": 33230 + }, + { + "epoch": 0.5185000311973544, + "grad_norm": 2.0594241619110107, + "learning_rate": 0.00016543437116532517, + "loss": 0.1885, + "step": 33240 + }, + { + "epoch": 0.5186560179696762, + "grad_norm": 0.5594693422317505, + "learning_rate": 0.00016542397204717042, + "loss": 0.4333, + "step": 33250 + }, + { + "epoch": 0.5188120047419978, + "grad_norm": 4.314854621887207, + "learning_rate": 0.00016541357292901561, + "loss": 0.2103, + "step": 33260 + }, + { + "epoch": 0.5189679915143196, + "grad_norm": 0.5609000325202942, + "learning_rate": 0.00016540317381086086, + "loss": 0.2498, + "step": 33270 + }, + { + "epoch": 0.5191239782866413, + "grad_norm": 3.680088758468628, + "learning_rate": 0.00016539277469270605, + "loss": 0.3465, + "step": 33280 + }, + { + "epoch": 0.519279965058963, + "grad_norm": 1.4993672370910645, + "learning_rate": 0.0001653823755745513, + "loss": 0.3272, + "step": 33290 + }, + { + "epoch": 0.5194359518312847, + "grad_norm": 0.18005084991455078, + "learning_rate": 0.0001653719764563965, + "loss": 0.1706, + "step": 33300 + }, + { + "epoch": 0.5195919386036064, + "grad_norm": 0.5105524659156799, + "learning_rate": 0.00016536157733824174, + "loss": 0.0937, + "step": 33310 + }, + { + "epoch": 0.5197479253759281, + "grad_norm": 1.2806589603424072, + "learning_rate": 0.00016535117822008693, + "loss": 0.1578, + "step": 33320 + }, + { + "epoch": 0.5199039121482498, + "grad_norm": 0.22015990316867828, + "learning_rate": 0.00016534077910193218, + "loss": 0.2081, + "step": 33330 + }, + { + "epoch": 0.5200598989205716, + "grad_norm": 0.34638991951942444, + "learning_rate": 0.00016533037998377737, + "loss": 0.2389, + "step": 33340 + }, + { + "epoch": 0.5202158856928932, + "grad_norm": 0.8709302544593811, + "learning_rate": 0.0001653199808656226, + "loss": 0.2112, + "step": 33350 + }, + { + "epoch": 0.520371872465215, + "grad_norm": 1.2309521436691284, + "learning_rate": 0.0001653095817474678, + "loss": 0.3448, + "step": 33360 + }, + { + "epoch": 0.5205278592375366, + "grad_norm": 1.3155115842819214, + "learning_rate": 0.00016529918262931303, + "loss": 0.1572, + "step": 33370 + }, + { + "epoch": 0.5206838460098584, + "grad_norm": 0.358751118183136, + "learning_rate": 0.00016528878351115825, + "loss": 0.3011, + "step": 33380 + }, + { + "epoch": 0.52083983278218, + "grad_norm": 3.800628662109375, + "learning_rate": 0.00016527838439300347, + "loss": 0.458, + "step": 33390 + }, + { + "epoch": 0.5209958195545018, + "grad_norm": 1.1253124475479126, + "learning_rate": 0.0001652679852748487, + "loss": 0.1238, + "step": 33400 + }, + { + "epoch": 0.5211518063268235, + "grad_norm": 0.46979108452796936, + "learning_rate": 0.0001652575861566939, + "loss": 0.0745, + "step": 33410 + }, + { + "epoch": 0.5213077930991452, + "grad_norm": 7.921492576599121, + "learning_rate": 0.00016524718703853913, + "loss": 0.4004, + "step": 33420 + }, + { + "epoch": 0.5214637798714669, + "grad_norm": 1.643127679824829, + "learning_rate": 0.00016523678792038435, + "loss": 0.2858, + "step": 33430 + }, + { + "epoch": 0.5216197666437886, + "grad_norm": 7.096631050109863, + "learning_rate": 0.0001652263888022296, + "loss": 0.4735, + "step": 33440 + }, + { + "epoch": 0.5217757534161103, + "grad_norm": 0.7280219197273254, + "learning_rate": 0.0001652159896840748, + "loss": 0.1841, + "step": 33450 + }, + { + "epoch": 0.521931740188432, + "grad_norm": 0.9415522813796997, + "learning_rate": 0.00016520559056592003, + "loss": 0.2169, + "step": 33460 + }, + { + "epoch": 0.5220877269607537, + "grad_norm": 2.264862060546875, + "learning_rate": 0.00016519519144776523, + "loss": 0.1902, + "step": 33470 + }, + { + "epoch": 0.5222437137330754, + "grad_norm": 2.4566822052001953, + "learning_rate": 0.00016518479232961047, + "loss": 0.3329, + "step": 33480 + }, + { + "epoch": 0.5223997005053972, + "grad_norm": 0.7330378890037537, + "learning_rate": 0.00016517439321145567, + "loss": 0.2341, + "step": 33490 + }, + { + "epoch": 0.5225556872777188, + "grad_norm": 1.6409542560577393, + "learning_rate": 0.0001651639940933009, + "loss": 0.3077, + "step": 33500 + }, + { + "epoch": 0.5227116740500406, + "grad_norm": 1.0647432804107666, + "learning_rate": 0.0001651535949751461, + "loss": 0.1935, + "step": 33510 + }, + { + "epoch": 0.5228676608223622, + "grad_norm": 0.507420539855957, + "learning_rate": 0.00016514319585699135, + "loss": 0.1314, + "step": 33520 + }, + { + "epoch": 0.523023647594684, + "grad_norm": 1.0064164400100708, + "learning_rate": 0.00016513279673883654, + "loss": 0.344, + "step": 33530 + }, + { + "epoch": 0.5231796343670057, + "grad_norm": 1.1936209201812744, + "learning_rate": 0.0001651223976206818, + "loss": 0.1948, + "step": 33540 + }, + { + "epoch": 0.5233356211393274, + "grad_norm": 0.04992926865816116, + "learning_rate": 0.00016511199850252698, + "loss": 0.1363, + "step": 33550 + }, + { + "epoch": 0.5234916079116491, + "grad_norm": 1.8256275653839111, + "learning_rate": 0.00016510159938437223, + "loss": 0.443, + "step": 33560 + }, + { + "epoch": 0.5236475946839708, + "grad_norm": 4.914908409118652, + "learning_rate": 0.00016509120026621742, + "loss": 0.4654, + "step": 33570 + }, + { + "epoch": 0.5238035814562925, + "grad_norm": 1.1625486612319946, + "learning_rate": 0.00016508080114806267, + "loss": 0.1974, + "step": 33580 + }, + { + "epoch": 0.5239595682286142, + "grad_norm": 6.6192522048950195, + "learning_rate": 0.00016507040202990786, + "loss": 0.2131, + "step": 33590 + }, + { + "epoch": 0.5241155550009359, + "grad_norm": 0.7369065880775452, + "learning_rate": 0.0001650600029117531, + "loss": 0.2371, + "step": 33600 + }, + { + "epoch": 0.5242715417732576, + "grad_norm": 1.5238152742385864, + "learning_rate": 0.0001650496037935983, + "loss": 0.1557, + "step": 33610 + }, + { + "epoch": 0.5244275285455793, + "grad_norm": 1.0418007373809814, + "learning_rate": 0.00016503920467544355, + "loss": 0.1878, + "step": 33620 + }, + { + "epoch": 0.524583515317901, + "grad_norm": 0.790117084980011, + "learning_rate": 0.00016502880555728874, + "loss": 0.2195, + "step": 33630 + }, + { + "epoch": 0.5247395020902228, + "grad_norm": 1.6712257862091064, + "learning_rate": 0.000165018406439134, + "loss": 0.1602, + "step": 33640 + }, + { + "epoch": 0.5248954888625444, + "grad_norm": 0.19236230850219727, + "learning_rate": 0.00016500800732097918, + "loss": 0.2526, + "step": 33650 + }, + { + "epoch": 0.5250514756348662, + "grad_norm": 1.3519701957702637, + "learning_rate": 0.00016499760820282443, + "loss": 0.4686, + "step": 33660 + }, + { + "epoch": 0.5252074624071879, + "grad_norm": 1.694342851638794, + "learning_rate": 0.00016498720908466962, + "loss": 0.1859, + "step": 33670 + }, + { + "epoch": 0.5253634491795096, + "grad_norm": 5.225239276885986, + "learning_rate": 0.00016497680996651487, + "loss": 0.2944, + "step": 33680 + }, + { + "epoch": 0.5255194359518313, + "grad_norm": 2.0208842754364014, + "learning_rate": 0.00016496641084836006, + "loss": 0.2421, + "step": 33690 + }, + { + "epoch": 0.525675422724153, + "grad_norm": 0.7954996228218079, + "learning_rate": 0.0001649560117302053, + "loss": 0.2853, + "step": 33700 + }, + { + "epoch": 0.5258314094964747, + "grad_norm": 2.296086072921753, + "learning_rate": 0.0001649456126120505, + "loss": 0.1333, + "step": 33710 + }, + { + "epoch": 0.5259873962687964, + "grad_norm": 1.1779128313064575, + "learning_rate": 0.00016493521349389574, + "loss": 0.1066, + "step": 33720 + }, + { + "epoch": 0.5261433830411181, + "grad_norm": 0.1756065934896469, + "learning_rate": 0.00016492481437574094, + "loss": 0.1352, + "step": 33730 + }, + { + "epoch": 0.5262993698134398, + "grad_norm": 0.13100725412368774, + "learning_rate": 0.00016491441525758618, + "loss": 0.2399, + "step": 33740 + }, + { + "epoch": 0.5264553565857615, + "grad_norm": 5.532008171081543, + "learning_rate": 0.00016490401613943138, + "loss": 0.2896, + "step": 33750 + }, + { + "epoch": 0.5266113433580832, + "grad_norm": 1.319886565208435, + "learning_rate": 0.00016489361702127662, + "loss": 0.3275, + "step": 33760 + }, + { + "epoch": 0.5267673301304049, + "grad_norm": 1.5550974607467651, + "learning_rate": 0.00016488321790312182, + "loss": 0.2677, + "step": 33770 + }, + { + "epoch": 0.5269233169027266, + "grad_norm": 1.8936737775802612, + "learning_rate": 0.00016487281878496706, + "loss": 0.1955, + "step": 33780 + }, + { + "epoch": 0.5270793036750484, + "grad_norm": 0.3653401732444763, + "learning_rate": 0.00016486241966681226, + "loss": 0.0723, + "step": 33790 + }, + { + "epoch": 0.52723529044737, + "grad_norm": 2.861341714859009, + "learning_rate": 0.00016485202054865747, + "loss": 0.2412, + "step": 33800 + }, + { + "epoch": 0.5273912772196918, + "grad_norm": 1.5291428565979004, + "learning_rate": 0.0001648416214305027, + "loss": 0.0871, + "step": 33810 + }, + { + "epoch": 0.5275472639920135, + "grad_norm": 1.0372581481933594, + "learning_rate": 0.00016483122231234791, + "loss": 0.4705, + "step": 33820 + }, + { + "epoch": 0.5277032507643352, + "grad_norm": 1.1943141222000122, + "learning_rate": 0.00016482082319419313, + "loss": 0.2848, + "step": 33830 + }, + { + "epoch": 0.5278592375366569, + "grad_norm": 1.9008225202560425, + "learning_rate": 0.00016481042407603835, + "loss": 0.139, + "step": 33840 + }, + { + "epoch": 0.5280152243089786, + "grad_norm": 2.132089138031006, + "learning_rate": 0.00016480002495788357, + "loss": 0.2119, + "step": 33850 + }, + { + "epoch": 0.5281712110813003, + "grad_norm": 0.24524426460266113, + "learning_rate": 0.0001647896258397288, + "loss": 0.1643, + "step": 33860 + }, + { + "epoch": 0.528327197853622, + "grad_norm": 1.6469637155532837, + "learning_rate": 0.000164779226721574, + "loss": 0.363, + "step": 33870 + }, + { + "epoch": 0.5284831846259437, + "grad_norm": 0.8767328858375549, + "learning_rate": 0.00016476882760341923, + "loss": 0.2632, + "step": 33880 + }, + { + "epoch": 0.5286391713982654, + "grad_norm": 0.06347586214542389, + "learning_rate": 0.00016475842848526445, + "loss": 0.3204, + "step": 33890 + }, + { + "epoch": 0.5287951581705871, + "grad_norm": 0.09782540798187256, + "learning_rate": 0.00016474802936710967, + "loss": 0.2211, + "step": 33900 + }, + { + "epoch": 0.5289511449429088, + "grad_norm": 3.2998859882354736, + "learning_rate": 0.0001647376302489549, + "loss": 0.265, + "step": 33910 + }, + { + "epoch": 0.5291071317152305, + "grad_norm": 0.43594226241111755, + "learning_rate": 0.0001647272311308001, + "loss": 0.194, + "step": 33920 + }, + { + "epoch": 0.5292631184875523, + "grad_norm": 1.5166605710983276, + "learning_rate": 0.00016471683201264533, + "loss": 0.2675, + "step": 33930 + }, + { + "epoch": 0.529419105259874, + "grad_norm": 0.6056640148162842, + "learning_rate": 0.00016470643289449055, + "loss": 0.0576, + "step": 33940 + }, + { + "epoch": 0.5295750920321957, + "grad_norm": 0.25410348176956177, + "learning_rate": 0.00016469603377633577, + "loss": 0.1124, + "step": 33950 + }, + { + "epoch": 0.5297310788045174, + "grad_norm": 1.770642876625061, + "learning_rate": 0.000164685634658181, + "loss": 0.3295, + "step": 33960 + }, + { + "epoch": 0.5298870655768391, + "grad_norm": 0.0607205331325531, + "learning_rate": 0.0001646752355400262, + "loss": 0.2369, + "step": 33970 + }, + { + "epoch": 0.5300430523491608, + "grad_norm": 0.5557095408439636, + "learning_rate": 0.00016466483642187143, + "loss": 0.1681, + "step": 33980 + }, + { + "epoch": 0.5301990391214825, + "grad_norm": 0.5192957520484924, + "learning_rate": 0.00016465443730371665, + "loss": 0.2649, + "step": 33990 + }, + { + "epoch": 0.5303550258938042, + "grad_norm": 0.04804835096001625, + "learning_rate": 0.00016464403818556187, + "loss": 0.2639, + "step": 34000 + }, + { + "epoch": 0.5305110126661259, + "grad_norm": 0.02673129364848137, + "learning_rate": 0.0001646336390674071, + "loss": 0.169, + "step": 34010 + }, + { + "epoch": 0.5306669994384476, + "grad_norm": 1.0084244012832642, + "learning_rate": 0.0001646232399492523, + "loss": 0.353, + "step": 34020 + }, + { + "epoch": 0.5308229862107693, + "grad_norm": 2.2202091217041016, + "learning_rate": 0.00016461284083109753, + "loss": 0.3206, + "step": 34030 + }, + { + "epoch": 0.530978972983091, + "grad_norm": 0.5573744773864746, + "learning_rate": 0.00016460244171294275, + "loss": 0.231, + "step": 34040 + }, + { + "epoch": 0.5311349597554127, + "grad_norm": 0.39700084924697876, + "learning_rate": 0.00016459204259478797, + "loss": 0.1412, + "step": 34050 + }, + { + "epoch": 0.5312909465277345, + "grad_norm": 2.582963228225708, + "learning_rate": 0.00016458164347663319, + "loss": 0.3334, + "step": 34060 + }, + { + "epoch": 0.5314469333000561, + "grad_norm": 0.22781169414520264, + "learning_rate": 0.0001645712443584784, + "loss": 0.2767, + "step": 34070 + }, + { + "epoch": 0.5316029200723779, + "grad_norm": 1.2051042318344116, + "learning_rate": 0.00016456084524032362, + "loss": 0.2045, + "step": 34080 + }, + { + "epoch": 0.5317589068446996, + "grad_norm": 0.42760100960731506, + "learning_rate": 0.00016455044612216884, + "loss": 0.1151, + "step": 34090 + }, + { + "epoch": 0.5319148936170213, + "grad_norm": 0.14440476894378662, + "learning_rate": 0.00016454004700401406, + "loss": 0.2386, + "step": 34100 + }, + { + "epoch": 0.532070880389343, + "grad_norm": 2.2777981758117676, + "learning_rate": 0.00016452964788585928, + "loss": 0.2009, + "step": 34110 + }, + { + "epoch": 0.5322268671616647, + "grad_norm": 0.9206979274749756, + "learning_rate": 0.0001645192487677045, + "loss": 0.2745, + "step": 34120 + }, + { + "epoch": 0.5323828539339864, + "grad_norm": 1.6947574615478516, + "learning_rate": 0.00016450884964954972, + "loss": 0.2584, + "step": 34130 + }, + { + "epoch": 0.5325388407063081, + "grad_norm": 0.401444673538208, + "learning_rate": 0.00016449845053139494, + "loss": 0.2218, + "step": 34140 + }, + { + "epoch": 0.5326948274786298, + "grad_norm": 0.08261553198099136, + "learning_rate": 0.00016448805141324016, + "loss": 0.2775, + "step": 34150 + }, + { + "epoch": 0.5328508142509515, + "grad_norm": 0.1017974391579628, + "learning_rate": 0.00016447765229508538, + "loss": 0.2095, + "step": 34160 + }, + { + "epoch": 0.5330068010232732, + "grad_norm": 1.3759571313858032, + "learning_rate": 0.0001644672531769306, + "loss": 0.2643, + "step": 34170 + }, + { + "epoch": 0.5331627877955949, + "grad_norm": 1.2654389142990112, + "learning_rate": 0.00016445685405877582, + "loss": 0.2949, + "step": 34180 + }, + { + "epoch": 0.5333187745679167, + "grad_norm": 1.5481843948364258, + "learning_rate": 0.00016444645494062104, + "loss": 0.1264, + "step": 34190 + }, + { + "epoch": 0.5334747613402383, + "grad_norm": 1.8094528913497925, + "learning_rate": 0.00016443605582246626, + "loss": 0.2727, + "step": 34200 + }, + { + "epoch": 0.5336307481125601, + "grad_norm": 2.224538564682007, + "learning_rate": 0.00016442565670431148, + "loss": 0.3096, + "step": 34210 + }, + { + "epoch": 0.5337867348848817, + "grad_norm": 0.6375226974487305, + "learning_rate": 0.0001644152575861567, + "loss": 0.2251, + "step": 34220 + }, + { + "epoch": 0.5339427216572035, + "grad_norm": 3.727106809616089, + "learning_rate": 0.00016440485846800192, + "loss": 0.4374, + "step": 34230 + }, + { + "epoch": 0.5340987084295252, + "grad_norm": 0.13345426321029663, + "learning_rate": 0.00016439445934984714, + "loss": 0.2011, + "step": 34240 + }, + { + "epoch": 0.5342546952018469, + "grad_norm": 2.1658668518066406, + "learning_rate": 0.00016438406023169236, + "loss": 0.2457, + "step": 34250 + }, + { + "epoch": 0.5344106819741686, + "grad_norm": 19.238407135009766, + "learning_rate": 0.00016437366111353758, + "loss": 0.2756, + "step": 34260 + }, + { + "epoch": 0.5345666687464903, + "grad_norm": 1.0292778015136719, + "learning_rate": 0.0001643632619953828, + "loss": 0.1646, + "step": 34270 + }, + { + "epoch": 0.534722655518812, + "grad_norm": 0.9372987747192383, + "learning_rate": 0.00016435286287722802, + "loss": 0.2762, + "step": 34280 + }, + { + "epoch": 0.5348786422911337, + "grad_norm": 0.3918002247810364, + "learning_rate": 0.00016434246375907324, + "loss": 0.1406, + "step": 34290 + }, + { + "epoch": 0.5350346290634554, + "grad_norm": 1.3518732786178589, + "learning_rate": 0.00016433206464091846, + "loss": 0.2992, + "step": 34300 + }, + { + "epoch": 0.5351906158357771, + "grad_norm": 0.73117595911026, + "learning_rate": 0.00016432166552276368, + "loss": 0.1172, + "step": 34310 + }, + { + "epoch": 0.5353466026080989, + "grad_norm": 0.06655958294868469, + "learning_rate": 0.0001643112664046089, + "loss": 0.1332, + "step": 34320 + }, + { + "epoch": 0.5355025893804205, + "grad_norm": 3.2587168216705322, + "learning_rate": 0.00016430086728645412, + "loss": 0.2052, + "step": 34330 + }, + { + "epoch": 0.5356585761527423, + "grad_norm": 2.6668431758880615, + "learning_rate": 0.00016429046816829934, + "loss": 0.1295, + "step": 34340 + }, + { + "epoch": 0.5358145629250639, + "grad_norm": 1.316080927848816, + "learning_rate": 0.00016428006905014456, + "loss": 0.2072, + "step": 34350 + }, + { + "epoch": 0.5359705496973857, + "grad_norm": 5.144528388977051, + "learning_rate": 0.00016426966993198977, + "loss": 0.1934, + "step": 34360 + }, + { + "epoch": 0.5361265364697073, + "grad_norm": 1.3529599905014038, + "learning_rate": 0.000164259270813835, + "loss": 0.2077, + "step": 34370 + }, + { + "epoch": 0.5362825232420291, + "grad_norm": 0.9132925868034363, + "learning_rate": 0.00016424887169568021, + "loss": 0.2404, + "step": 34380 + }, + { + "epoch": 0.5364385100143508, + "grad_norm": 1.520033597946167, + "learning_rate": 0.00016423847257752543, + "loss": 0.1765, + "step": 34390 + }, + { + "epoch": 0.5365944967866725, + "grad_norm": 2.343975305557251, + "learning_rate": 0.00016422807345937065, + "loss": 0.1959, + "step": 34400 + }, + { + "epoch": 0.5367504835589942, + "grad_norm": 0.5855118632316589, + "learning_rate": 0.00016421767434121587, + "loss": 0.0848, + "step": 34410 + }, + { + "epoch": 0.5369064703313159, + "grad_norm": 1.9108648300170898, + "learning_rate": 0.0001642072752230611, + "loss": 0.2406, + "step": 34420 + }, + { + "epoch": 0.5370624571036376, + "grad_norm": 1.0323792695999146, + "learning_rate": 0.0001641968761049063, + "loss": 0.2298, + "step": 34430 + }, + { + "epoch": 0.5372184438759593, + "grad_norm": 0.17435620725154877, + "learning_rate": 0.00016418647698675153, + "loss": 0.1396, + "step": 34440 + }, + { + "epoch": 0.537374430648281, + "grad_norm": 0.3526577353477478, + "learning_rate": 0.00016417607786859675, + "loss": 0.172, + "step": 34450 + }, + { + "epoch": 0.5375304174206027, + "grad_norm": 2.73685884475708, + "learning_rate": 0.00016416567875044197, + "loss": 0.1167, + "step": 34460 + }, + { + "epoch": 0.5376864041929245, + "grad_norm": 2.765693426132202, + "learning_rate": 0.0001641552796322872, + "loss": 0.1356, + "step": 34470 + }, + { + "epoch": 0.5378423909652461, + "grad_norm": 0.47628021240234375, + "learning_rate": 0.0001641448805141324, + "loss": 0.3104, + "step": 34480 + }, + { + "epoch": 0.5379983777375679, + "grad_norm": 1.1307590007781982, + "learning_rate": 0.00016413448139597763, + "loss": 0.3247, + "step": 34490 + }, + { + "epoch": 0.5381543645098895, + "grad_norm": 0.07136381417512894, + "learning_rate": 0.00016412408227782285, + "loss": 0.319, + "step": 34500 + }, + { + "epoch": 0.5383103512822113, + "grad_norm": 1.7850221395492554, + "learning_rate": 0.00016411368315966807, + "loss": 0.2515, + "step": 34510 + }, + { + "epoch": 0.5384663380545329, + "grad_norm": 1.3870742321014404, + "learning_rate": 0.0001641032840415133, + "loss": 0.256, + "step": 34520 + }, + { + "epoch": 0.5386223248268547, + "grad_norm": 1.1329221725463867, + "learning_rate": 0.0001640928849233585, + "loss": 0.2255, + "step": 34530 + }, + { + "epoch": 0.5387783115991764, + "grad_norm": 0.23236137628555298, + "learning_rate": 0.00016408248580520373, + "loss": 0.4053, + "step": 34540 + }, + { + "epoch": 0.5389342983714981, + "grad_norm": 3.350924253463745, + "learning_rate": 0.00016407208668704895, + "loss": 0.2816, + "step": 34550 + }, + { + "epoch": 0.5390902851438198, + "grad_norm": 1.7459170818328857, + "learning_rate": 0.00016406168756889417, + "loss": 0.0834, + "step": 34560 + }, + { + "epoch": 0.5392462719161415, + "grad_norm": 0.89351487159729, + "learning_rate": 0.0001640512884507394, + "loss": 0.1146, + "step": 34570 + }, + { + "epoch": 0.5394022586884633, + "grad_norm": 0.49410831928253174, + "learning_rate": 0.0001640408893325846, + "loss": 0.2319, + "step": 34580 + }, + { + "epoch": 0.5395582454607849, + "grad_norm": 0.009971237741410732, + "learning_rate": 0.00016403049021442983, + "loss": 0.1674, + "step": 34590 + }, + { + "epoch": 0.5397142322331067, + "grad_norm": 1.420511245727539, + "learning_rate": 0.00016402009109627505, + "loss": 0.1469, + "step": 34600 + }, + { + "epoch": 0.5398702190054283, + "grad_norm": 2.5686769485473633, + "learning_rate": 0.00016400969197812027, + "loss": 0.2607, + "step": 34610 + }, + { + "epoch": 0.5400262057777501, + "grad_norm": 2.5808913707733154, + "learning_rate": 0.00016399929285996549, + "loss": 0.4412, + "step": 34620 + }, + { + "epoch": 0.5401821925500717, + "grad_norm": 0.2778591811656952, + "learning_rate": 0.0001639888937418107, + "loss": 0.1563, + "step": 34630 + }, + { + "epoch": 0.5403381793223935, + "grad_norm": 1.6448099613189697, + "learning_rate": 0.00016397849462365592, + "loss": 0.2484, + "step": 34640 + }, + { + "epoch": 0.5404941660947151, + "grad_norm": 0.003213417250663042, + "learning_rate": 0.00016396809550550114, + "loss": 0.1582, + "step": 34650 + }, + { + "epoch": 0.5406501528670369, + "grad_norm": 2.3433990478515625, + "learning_rate": 0.00016395769638734636, + "loss": 0.2774, + "step": 34660 + }, + { + "epoch": 0.5408061396393585, + "grad_norm": 0.09544426947832108, + "learning_rate": 0.00016394729726919158, + "loss": 0.1242, + "step": 34670 + }, + { + "epoch": 0.5409621264116803, + "grad_norm": 1.8934235572814941, + "learning_rate": 0.0001639368981510368, + "loss": 0.1636, + "step": 34680 + }, + { + "epoch": 0.541118113184002, + "grad_norm": 2.8528554439544678, + "learning_rate": 0.00016392649903288202, + "loss": 0.2524, + "step": 34690 + }, + { + "epoch": 0.5412740999563237, + "grad_norm": 2.837836980819702, + "learning_rate": 0.00016391609991472722, + "loss": 0.1808, + "step": 34700 + }, + { + "epoch": 0.5414300867286455, + "grad_norm": 4.349740028381348, + "learning_rate": 0.00016390570079657246, + "loss": 0.12, + "step": 34710 + }, + { + "epoch": 0.5415860735009671, + "grad_norm": 0.6011945605278015, + "learning_rate": 0.00016389530167841765, + "loss": 0.2865, + "step": 34720 + }, + { + "epoch": 0.5417420602732889, + "grad_norm": 1.2079488039016724, + "learning_rate": 0.0001638849025602629, + "loss": 0.125, + "step": 34730 + }, + { + "epoch": 0.5418980470456105, + "grad_norm": 1.9549164772033691, + "learning_rate": 0.0001638745034421081, + "loss": 0.3285, + "step": 34740 + }, + { + "epoch": 0.5420540338179323, + "grad_norm": 0.6969407200813293, + "learning_rate": 0.00016386410432395334, + "loss": 0.3267, + "step": 34750 + }, + { + "epoch": 0.5422100205902539, + "grad_norm": 1.0561965703964233, + "learning_rate": 0.00016385370520579853, + "loss": 0.1463, + "step": 34760 + }, + { + "epoch": 0.5423660073625757, + "grad_norm": 1.8208931684494019, + "learning_rate": 0.00016384330608764378, + "loss": 0.1713, + "step": 34770 + }, + { + "epoch": 0.5425219941348973, + "grad_norm": 0.6349910497665405, + "learning_rate": 0.000163832906969489, + "loss": 0.2244, + "step": 34780 + }, + { + "epoch": 0.5426779809072191, + "grad_norm": 1.1976515054702759, + "learning_rate": 0.00016382250785133422, + "loss": 0.3407, + "step": 34790 + }, + { + "epoch": 0.5428339676795407, + "grad_norm": 0.5970319509506226, + "learning_rate": 0.00016381210873317944, + "loss": 0.1419, + "step": 34800 + }, + { + "epoch": 0.5429899544518625, + "grad_norm": 0.14623159170150757, + "learning_rate": 0.00016380170961502466, + "loss": 0.14, + "step": 34810 + }, + { + "epoch": 0.5431459412241841, + "grad_norm": 1.4332351684570312, + "learning_rate": 0.00016379131049686988, + "loss": 0.1093, + "step": 34820 + }, + { + "epoch": 0.5433019279965059, + "grad_norm": 1.1443063020706177, + "learning_rate": 0.0001637809113787151, + "loss": 0.3676, + "step": 34830 + }, + { + "epoch": 0.5434579147688277, + "grad_norm": 1.3897461891174316, + "learning_rate": 0.00016377051226056032, + "loss": 0.3713, + "step": 34840 + }, + { + "epoch": 0.5436139015411493, + "grad_norm": 1.7773199081420898, + "learning_rate": 0.00016376011314240554, + "loss": 0.2249, + "step": 34850 + }, + { + "epoch": 0.5437698883134711, + "grad_norm": 0.8524389266967773, + "learning_rate": 0.00016374971402425076, + "loss": 0.2158, + "step": 34860 + }, + { + "epoch": 0.5439258750857927, + "grad_norm": 0.5572389960289001, + "learning_rate": 0.00016373931490609598, + "loss": 0.215, + "step": 34870 + }, + { + "epoch": 0.5440818618581145, + "grad_norm": 3.0238709449768066, + "learning_rate": 0.0001637289157879412, + "loss": 0.3765, + "step": 34880 + }, + { + "epoch": 0.5442378486304361, + "grad_norm": 1.1655421257019043, + "learning_rate": 0.00016371851666978642, + "loss": 0.2491, + "step": 34890 + }, + { + "epoch": 0.5443938354027579, + "grad_norm": 0.49492090940475464, + "learning_rate": 0.00016370811755163164, + "loss": 0.1287, + "step": 34900 + }, + { + "epoch": 0.5445498221750795, + "grad_norm": 0.2732921540737152, + "learning_rate": 0.00016369771843347685, + "loss": 0.2154, + "step": 34910 + }, + { + "epoch": 0.5447058089474013, + "grad_norm": 2.5807769298553467, + "learning_rate": 0.00016368731931532207, + "loss": 0.2032, + "step": 34920 + }, + { + "epoch": 0.5448617957197229, + "grad_norm": 0.5044315457344055, + "learning_rate": 0.0001636769201971673, + "loss": 0.2595, + "step": 34930 + }, + { + "epoch": 0.5450177824920447, + "grad_norm": 2.1248972415924072, + "learning_rate": 0.00016366652107901251, + "loss": 0.5651, + "step": 34940 + }, + { + "epoch": 0.5451737692643663, + "grad_norm": 0.8391468524932861, + "learning_rate": 0.00016365612196085773, + "loss": 0.3028, + "step": 34950 + }, + { + "epoch": 0.5453297560366881, + "grad_norm": 0.36081477999687195, + "learning_rate": 0.00016364572284270295, + "loss": 0.1073, + "step": 34960 + }, + { + "epoch": 0.5454857428090097, + "grad_norm": 3.175804376602173, + "learning_rate": 0.00016363532372454817, + "loss": 0.1889, + "step": 34970 + }, + { + "epoch": 0.5456417295813315, + "grad_norm": 0.20647937059402466, + "learning_rate": 0.0001636249246063934, + "loss": 0.2416, + "step": 34980 + }, + { + "epoch": 0.5457977163536533, + "grad_norm": 2.3525943756103516, + "learning_rate": 0.0001636145254882386, + "loss": 0.3179, + "step": 34990 + }, + { + "epoch": 0.5459537031259749, + "grad_norm": 0.2641216814517975, + "learning_rate": 0.00016360412637008383, + "loss": 0.2157, + "step": 35000 + }, + { + "epoch": 0.5461096898982967, + "grad_norm": 0.09906073659658432, + "learning_rate": 0.00016359372725192905, + "loss": 0.2576, + "step": 35010 + }, + { + "epoch": 0.5462656766706183, + "grad_norm": 0.13971920311450958, + "learning_rate": 0.00016358332813377427, + "loss": 0.1298, + "step": 35020 + }, + { + "epoch": 0.5464216634429401, + "grad_norm": 1.4535638093948364, + "learning_rate": 0.0001635729290156195, + "loss": 0.2357, + "step": 35030 + }, + { + "epoch": 0.5465776502152617, + "grad_norm": 0.9107828736305237, + "learning_rate": 0.0001635625298974647, + "loss": 0.3049, + "step": 35040 + }, + { + "epoch": 0.5467336369875835, + "grad_norm": 0.0377386212348938, + "learning_rate": 0.00016355213077930993, + "loss": 0.2066, + "step": 35050 + }, + { + "epoch": 0.5468896237599051, + "grad_norm": 2.7800869941711426, + "learning_rate": 0.00016354173166115515, + "loss": 0.2974, + "step": 35060 + }, + { + "epoch": 0.5470456105322269, + "grad_norm": 1.6247998476028442, + "learning_rate": 0.00016353133254300037, + "loss": 0.3822, + "step": 35070 + }, + { + "epoch": 0.5472015973045485, + "grad_norm": 1.6479015350341797, + "learning_rate": 0.0001635209334248456, + "loss": 0.3401, + "step": 35080 + }, + { + "epoch": 0.5473575840768703, + "grad_norm": 0.1879737675189972, + "learning_rate": 0.0001635105343066908, + "loss": 0.1549, + "step": 35090 + }, + { + "epoch": 0.5475135708491919, + "grad_norm": 0.39355361461639404, + "learning_rate": 0.00016350013518853603, + "loss": 0.2032, + "step": 35100 + }, + { + "epoch": 0.5476695576215137, + "grad_norm": 2.9976983070373535, + "learning_rate": 0.00016348973607038125, + "loss": 0.3867, + "step": 35110 + }, + { + "epoch": 0.5478255443938354, + "grad_norm": 0.8242707252502441, + "learning_rate": 0.00016347933695222647, + "loss": 0.3601, + "step": 35120 + }, + { + "epoch": 0.5479815311661571, + "grad_norm": 0.874955415725708, + "learning_rate": 0.0001634689378340717, + "loss": 0.2709, + "step": 35130 + }, + { + "epoch": 0.5481375179384789, + "grad_norm": 0.05980971083045006, + "learning_rate": 0.0001634585387159169, + "loss": 0.2646, + "step": 35140 + }, + { + "epoch": 0.5482935047108005, + "grad_norm": 1.8437328338623047, + "learning_rate": 0.0001634481395977621, + "loss": 0.1965, + "step": 35150 + }, + { + "epoch": 0.5484494914831223, + "grad_norm": 0.4142405688762665, + "learning_rate": 0.00016343774047960735, + "loss": 0.1756, + "step": 35160 + }, + { + "epoch": 0.5486054782554439, + "grad_norm": 4.836324214935303, + "learning_rate": 0.00016342734136145254, + "loss": 0.2534, + "step": 35170 + }, + { + "epoch": 0.5487614650277657, + "grad_norm": 1.4933065176010132, + "learning_rate": 0.00016341694224329779, + "loss": 0.2355, + "step": 35180 + }, + { + "epoch": 0.5489174518000873, + "grad_norm": 1.9468894004821777, + "learning_rate": 0.00016340654312514298, + "loss": 0.3113, + "step": 35190 + }, + { + "epoch": 0.5490734385724091, + "grad_norm": 1.012710452079773, + "learning_rate": 0.00016339614400698822, + "loss": 0.2635, + "step": 35200 + }, + { + "epoch": 0.5492294253447307, + "grad_norm": 1.5551140308380127, + "learning_rate": 0.00016338574488883342, + "loss": 0.2951, + "step": 35210 + }, + { + "epoch": 0.5493854121170525, + "grad_norm": 1.8613696098327637, + "learning_rate": 0.00016337534577067866, + "loss": 0.2832, + "step": 35220 + }, + { + "epoch": 0.5495413988893741, + "grad_norm": 1.6748839616775513, + "learning_rate": 0.00016336494665252386, + "loss": 0.2266, + "step": 35230 + }, + { + "epoch": 0.5496973856616959, + "grad_norm": 0.4094032347202301, + "learning_rate": 0.0001633545475343691, + "loss": 0.3497, + "step": 35240 + }, + { + "epoch": 0.5498533724340176, + "grad_norm": 1.6414631605148315, + "learning_rate": 0.0001633441484162143, + "loss": 0.1763, + "step": 35250 + }, + { + "epoch": 0.5500093592063393, + "grad_norm": 1.6380645036697388, + "learning_rate": 0.00016333374929805954, + "loss": 0.1355, + "step": 35260 + }, + { + "epoch": 0.550165345978661, + "grad_norm": 0.7325630187988281, + "learning_rate": 0.00016332335017990473, + "loss": 0.3105, + "step": 35270 + }, + { + "epoch": 0.5503213327509827, + "grad_norm": 1.6015644073486328, + "learning_rate": 0.00016331295106174998, + "loss": 0.2311, + "step": 35280 + }, + { + "epoch": 0.5504773195233045, + "grad_norm": 0.9224210381507874, + "learning_rate": 0.00016330255194359517, + "loss": 0.2746, + "step": 35290 + }, + { + "epoch": 0.5506333062956261, + "grad_norm": 1.30025315284729, + "learning_rate": 0.00016329215282544042, + "loss": 0.2066, + "step": 35300 + }, + { + "epoch": 0.5507892930679479, + "grad_norm": 0.481125146150589, + "learning_rate": 0.0001632817537072856, + "loss": 0.1778, + "step": 35310 + }, + { + "epoch": 0.5509452798402695, + "grad_norm": 0.2709486782550812, + "learning_rate": 0.00016327135458913086, + "loss": 0.1131, + "step": 35320 + }, + { + "epoch": 0.5511012666125913, + "grad_norm": 0.6248563528060913, + "learning_rate": 0.00016326095547097605, + "loss": 0.4356, + "step": 35330 + }, + { + "epoch": 0.5512572533849129, + "grad_norm": 3.0947647094726562, + "learning_rate": 0.0001632505563528213, + "loss": 0.3072, + "step": 35340 + }, + { + "epoch": 0.5514132401572347, + "grad_norm": 1.009535312652588, + "learning_rate": 0.0001632401572346665, + "loss": 0.1808, + "step": 35350 + }, + { + "epoch": 0.5515692269295563, + "grad_norm": 0.042604975402355194, + "learning_rate": 0.00016322975811651174, + "loss": 0.1198, + "step": 35360 + }, + { + "epoch": 0.5517252137018781, + "grad_norm": 1.8652396202087402, + "learning_rate": 0.00016321935899835693, + "loss": 0.2176, + "step": 35370 + }, + { + "epoch": 0.5518812004741998, + "grad_norm": 4.6097187995910645, + "learning_rate": 0.00016320895988020218, + "loss": 0.2132, + "step": 35380 + }, + { + "epoch": 0.5520371872465215, + "grad_norm": 1.3615522384643555, + "learning_rate": 0.00016319856076204737, + "loss": 0.2751, + "step": 35390 + }, + { + "epoch": 0.5521931740188432, + "grad_norm": 0.8844773769378662, + "learning_rate": 0.00016318816164389262, + "loss": 0.22, + "step": 35400 + }, + { + "epoch": 0.5523491607911649, + "grad_norm": 0.5413331985473633, + "learning_rate": 0.0001631777625257378, + "loss": 0.2159, + "step": 35410 + }, + { + "epoch": 0.5525051475634866, + "grad_norm": 1.5456678867340088, + "learning_rate": 0.00016316736340758306, + "loss": 0.2244, + "step": 35420 + }, + { + "epoch": 0.5526611343358083, + "grad_norm": 2.1405861377716064, + "learning_rate": 0.00016315696428942825, + "loss": 0.446, + "step": 35430 + }, + { + "epoch": 0.5528171211081301, + "grad_norm": 1.4269858598709106, + "learning_rate": 0.0001631465651712735, + "loss": 0.3599, + "step": 35440 + }, + { + "epoch": 0.5529731078804517, + "grad_norm": 1.8682516813278198, + "learning_rate": 0.0001631361660531187, + "loss": 0.115, + "step": 35450 + }, + { + "epoch": 0.5531290946527735, + "grad_norm": 1.0175774097442627, + "learning_rate": 0.00016312576693496394, + "loss": 0.4003, + "step": 35460 + }, + { + "epoch": 0.5532850814250951, + "grad_norm": 2.4369170665740967, + "learning_rate": 0.00016311536781680913, + "loss": 0.2489, + "step": 35470 + }, + { + "epoch": 0.5534410681974169, + "grad_norm": 1.886022686958313, + "learning_rate": 0.00016310496869865437, + "loss": 0.4201, + "step": 35480 + }, + { + "epoch": 0.5535970549697385, + "grad_norm": 1.5811959505081177, + "learning_rate": 0.00016309456958049957, + "loss": 0.4286, + "step": 35490 + }, + { + "epoch": 0.5537530417420603, + "grad_norm": 1.971110224723816, + "learning_rate": 0.00016308417046234481, + "loss": 0.3599, + "step": 35500 + }, + { + "epoch": 0.553909028514382, + "grad_norm": 0.0335380844771862, + "learning_rate": 0.00016307377134419, + "loss": 0.3357, + "step": 35510 + }, + { + "epoch": 0.5540650152867037, + "grad_norm": 1.8519577980041504, + "learning_rate": 0.00016306337222603525, + "loss": 0.3126, + "step": 35520 + }, + { + "epoch": 0.5542210020590254, + "grad_norm": 2.1563379764556885, + "learning_rate": 0.00016305297310788045, + "loss": 0.1123, + "step": 35530 + }, + { + "epoch": 0.5543769888313471, + "grad_norm": 1.64332914352417, + "learning_rate": 0.0001630425739897257, + "loss": 0.2853, + "step": 35540 + }, + { + "epoch": 0.5545329756036688, + "grad_norm": 0.061150554567575455, + "learning_rate": 0.00016303217487157088, + "loss": 0.1943, + "step": 35550 + }, + { + "epoch": 0.5546889623759905, + "grad_norm": 1.2701060771942139, + "learning_rate": 0.00016302177575341613, + "loss": 0.2304, + "step": 35560 + }, + { + "epoch": 0.5548449491483122, + "grad_norm": 2.424860715866089, + "learning_rate": 0.00016301137663526132, + "loss": 0.2128, + "step": 35570 + }, + { + "epoch": 0.5550009359206339, + "grad_norm": 0.6803575158119202, + "learning_rate": 0.00016300097751710657, + "loss": 0.1145, + "step": 35580 + }, + { + "epoch": 0.5551569226929557, + "grad_norm": 1.2855092287063599, + "learning_rate": 0.00016299057839895176, + "loss": 0.2448, + "step": 35590 + }, + { + "epoch": 0.5553129094652773, + "grad_norm": 2.6340911388397217, + "learning_rate": 0.00016298017928079698, + "loss": 0.1856, + "step": 35600 + }, + { + "epoch": 0.5554688962375991, + "grad_norm": 1.776382327079773, + "learning_rate": 0.0001629697801626422, + "loss": 0.2422, + "step": 35610 + }, + { + "epoch": 0.5556248830099207, + "grad_norm": 3.0746347904205322, + "learning_rate": 0.00016295938104448742, + "loss": 0.1578, + "step": 35620 + }, + { + "epoch": 0.5557808697822425, + "grad_norm": 1.4887659549713135, + "learning_rate": 0.00016294898192633267, + "loss": 0.2385, + "step": 35630 + }, + { + "epoch": 0.5559368565545642, + "grad_norm": 0.5371220707893372, + "learning_rate": 0.00016293858280817786, + "loss": 0.2361, + "step": 35640 + }, + { + "epoch": 0.5560928433268859, + "grad_norm": 0.18604904413223267, + "learning_rate": 0.0001629281836900231, + "loss": 0.0828, + "step": 35650 + }, + { + "epoch": 0.5562488300992076, + "grad_norm": 2.267854690551758, + "learning_rate": 0.0001629177845718683, + "loss": 0.413, + "step": 35660 + }, + { + "epoch": 0.5564048168715293, + "grad_norm": 1.9755452871322632, + "learning_rate": 0.00016290738545371355, + "loss": 0.146, + "step": 35670 + }, + { + "epoch": 0.556560803643851, + "grad_norm": 2.430293321609497, + "learning_rate": 0.00016289698633555874, + "loss": 0.2677, + "step": 35680 + }, + { + "epoch": 0.5567167904161727, + "grad_norm": 0.6319543719291687, + "learning_rate": 0.000162886587217404, + "loss": 0.1481, + "step": 35690 + }, + { + "epoch": 0.5568727771884944, + "grad_norm": 0.17439277470111847, + "learning_rate": 0.00016287618809924918, + "loss": 0.1641, + "step": 35700 + }, + { + "epoch": 0.5570287639608161, + "grad_norm": 0.350175142288208, + "learning_rate": 0.00016286578898109443, + "loss": 0.393, + "step": 35710 + }, + { + "epoch": 0.5571847507331378, + "grad_norm": 2.641941547393799, + "learning_rate": 0.00016285538986293962, + "loss": 0.1907, + "step": 35720 + }, + { + "epoch": 0.5573407375054595, + "grad_norm": 0.1910303384065628, + "learning_rate": 0.00016284499074478487, + "loss": 0.2248, + "step": 35730 + }, + { + "epoch": 0.5574967242777813, + "grad_norm": 1.0180896520614624, + "learning_rate": 0.00016283459162663006, + "loss": 0.1924, + "step": 35740 + }, + { + "epoch": 0.557652711050103, + "grad_norm": 1.2047260999679565, + "learning_rate": 0.0001628241925084753, + "loss": 0.2031, + "step": 35750 + }, + { + "epoch": 0.5578086978224247, + "grad_norm": 1.8182405233383179, + "learning_rate": 0.0001628137933903205, + "loss": 0.1573, + "step": 35760 + }, + { + "epoch": 0.5579646845947464, + "grad_norm": 0.5485963821411133, + "learning_rate": 0.00016280339427216574, + "loss": 0.152, + "step": 35770 + }, + { + "epoch": 0.5581206713670681, + "grad_norm": 0.1674145758152008, + "learning_rate": 0.00016279299515401094, + "loss": 0.2388, + "step": 35780 + }, + { + "epoch": 0.5582766581393898, + "grad_norm": 0.36187657713890076, + "learning_rate": 0.00016278259603585618, + "loss": 0.1618, + "step": 35790 + }, + { + "epoch": 0.5584326449117115, + "grad_norm": 0.013768521137535572, + "learning_rate": 0.00016277219691770138, + "loss": 0.3666, + "step": 35800 + }, + { + "epoch": 0.5585886316840332, + "grad_norm": 2.8602802753448486, + "learning_rate": 0.00016276179779954662, + "loss": 0.2065, + "step": 35810 + }, + { + "epoch": 0.5587446184563549, + "grad_norm": 1.7925455570220947, + "learning_rate": 0.00016275139868139182, + "loss": 0.1729, + "step": 35820 + }, + { + "epoch": 0.5589006052286766, + "grad_norm": 2.910456418991089, + "learning_rate": 0.00016274099956323706, + "loss": 0.3734, + "step": 35830 + }, + { + "epoch": 0.5590565920009983, + "grad_norm": 0.8378308415412903, + "learning_rate": 0.00016273060044508225, + "loss": 0.2255, + "step": 35840 + }, + { + "epoch": 0.55921257877332, + "grad_norm": 0.409534752368927, + "learning_rate": 0.0001627202013269275, + "loss": 0.2046, + "step": 35850 + }, + { + "epoch": 0.5593685655456417, + "grad_norm": 0.49498099088668823, + "learning_rate": 0.0001627098022087727, + "loss": 0.1058, + "step": 35860 + }, + { + "epoch": 0.5595245523179634, + "grad_norm": 1.1761783361434937, + "learning_rate": 0.00016269940309061794, + "loss": 0.1566, + "step": 35870 + }, + { + "epoch": 0.5596805390902851, + "grad_norm": 3.8204751014709473, + "learning_rate": 0.00016268900397246313, + "loss": 0.2646, + "step": 35880 + }, + { + "epoch": 0.5598365258626069, + "grad_norm": 0.9882522225379944, + "learning_rate": 0.00016267860485430838, + "loss": 0.1756, + "step": 35890 + }, + { + "epoch": 0.5599925126349286, + "grad_norm": 1.1832259893417358, + "learning_rate": 0.00016266820573615357, + "loss": 0.1385, + "step": 35900 + }, + { + "epoch": 0.5601484994072503, + "grad_norm": 0.7638296484947205, + "learning_rate": 0.00016265780661799882, + "loss": 0.5679, + "step": 35910 + }, + { + "epoch": 0.560304486179572, + "grad_norm": 0.8551504611968994, + "learning_rate": 0.000162647407499844, + "loss": 0.2981, + "step": 35920 + }, + { + "epoch": 0.5604604729518937, + "grad_norm": 5.499948501586914, + "learning_rate": 0.00016263700838168926, + "loss": 0.1429, + "step": 35930 + }, + { + "epoch": 0.5606164597242154, + "grad_norm": 3.0494496822357178, + "learning_rate": 0.00016262660926353445, + "loss": 0.1322, + "step": 35940 + }, + { + "epoch": 0.5607724464965371, + "grad_norm": 0.2482384592294693, + "learning_rate": 0.0001626162101453797, + "loss": 0.2473, + "step": 35950 + }, + { + "epoch": 0.5609284332688588, + "grad_norm": 0.009750776924192905, + "learning_rate": 0.0001626058110272249, + "loss": 0.0429, + "step": 35960 + }, + { + "epoch": 0.5610844200411805, + "grad_norm": 0.5742604732513428, + "learning_rate": 0.00016259541190907014, + "loss": 0.4565, + "step": 35970 + }, + { + "epoch": 0.5612404068135022, + "grad_norm": 0.7810243368148804, + "learning_rate": 0.00016258501279091533, + "loss": 0.2107, + "step": 35980 + }, + { + "epoch": 0.5613963935858239, + "grad_norm": 2.785747528076172, + "learning_rate": 0.00016257461367276058, + "loss": 0.5087, + "step": 35990 + }, + { + "epoch": 0.5615523803581456, + "grad_norm": 0.8806902766227722, + "learning_rate": 0.00016256421455460577, + "loss": 0.2937, + "step": 36000 + }, + { + "epoch": 0.5617083671304673, + "grad_norm": 1.857373833656311, + "learning_rate": 0.00016255381543645102, + "loss": 0.5042, + "step": 36010 + }, + { + "epoch": 0.561864353902789, + "grad_norm": 0.5333901047706604, + "learning_rate": 0.0001625434163182962, + "loss": 0.1678, + "step": 36020 + }, + { + "epoch": 0.5620203406751108, + "grad_norm": 11.023160934448242, + "learning_rate": 0.00016253301720014145, + "loss": 0.2721, + "step": 36030 + }, + { + "epoch": 0.5621763274474325, + "grad_norm": 2.6214029788970947, + "learning_rate": 0.00016252261808198665, + "loss": 0.1675, + "step": 36040 + }, + { + "epoch": 0.5623323142197542, + "grad_norm": 0.25980064272880554, + "learning_rate": 0.0001625122189638319, + "loss": 0.1832, + "step": 36050 + }, + { + "epoch": 0.5624883009920759, + "grad_norm": 1.3559473752975464, + "learning_rate": 0.0001625018198456771, + "loss": 0.121, + "step": 36060 + }, + { + "epoch": 0.5626442877643976, + "grad_norm": 0.3558153212070465, + "learning_rate": 0.0001624914207275223, + "loss": 0.2797, + "step": 36070 + }, + { + "epoch": 0.5628002745367193, + "grad_norm": 0.6314427256584167, + "learning_rate": 0.00016248102160936753, + "loss": 0.3455, + "step": 36080 + }, + { + "epoch": 0.562956261309041, + "grad_norm": 0.16120101511478424, + "learning_rate": 0.00016247062249121275, + "loss": 0.3669, + "step": 36090 + }, + { + "epoch": 0.5631122480813627, + "grad_norm": 1.041572093963623, + "learning_rate": 0.00016246022337305797, + "loss": 0.2744, + "step": 36100 + }, + { + "epoch": 0.5632682348536844, + "grad_norm": 0.2645890414714813, + "learning_rate": 0.00016244982425490318, + "loss": 0.1487, + "step": 36110 + }, + { + "epoch": 0.5634242216260061, + "grad_norm": 5.698698043823242, + "learning_rate": 0.0001624394251367484, + "loss": 0.173, + "step": 36120 + }, + { + "epoch": 0.5635802083983278, + "grad_norm": 0.3299804627895355, + "learning_rate": 0.00016242902601859362, + "loss": 0.1965, + "step": 36130 + }, + { + "epoch": 0.5637361951706495, + "grad_norm": 1.2476481199264526, + "learning_rate": 0.00016241862690043884, + "loss": 0.4097, + "step": 36140 + }, + { + "epoch": 0.5638921819429712, + "grad_norm": 0.30926263332366943, + "learning_rate": 0.00016240822778228406, + "loss": 0.1416, + "step": 36150 + }, + { + "epoch": 0.564048168715293, + "grad_norm": 0.05872740224003792, + "learning_rate": 0.00016239782866412928, + "loss": 0.1449, + "step": 36160 + }, + { + "epoch": 0.5642041554876146, + "grad_norm": 0.2292211800813675, + "learning_rate": 0.0001623874295459745, + "loss": 0.2557, + "step": 36170 + }, + { + "epoch": 0.5643601422599364, + "grad_norm": 1.7822531461715698, + "learning_rate": 0.00016237703042781972, + "loss": 0.2953, + "step": 36180 + }, + { + "epoch": 0.5645161290322581, + "grad_norm": 2.4908461570739746, + "learning_rate": 0.00016236663130966494, + "loss": 0.3888, + "step": 36190 + }, + { + "epoch": 0.5646721158045798, + "grad_norm": 0.08363594114780426, + "learning_rate": 0.00016235623219151016, + "loss": 0.2293, + "step": 36200 + }, + { + "epoch": 0.5648281025769015, + "grad_norm": 1.4830002784729004, + "learning_rate": 0.00016234583307335538, + "loss": 0.2348, + "step": 36210 + }, + { + "epoch": 0.5649840893492232, + "grad_norm": 1.8443071842193604, + "learning_rate": 0.0001623354339552006, + "loss": 0.3392, + "step": 36220 + }, + { + "epoch": 0.5651400761215449, + "grad_norm": 2.8050875663757324, + "learning_rate": 0.00016232503483704582, + "loss": 0.2426, + "step": 36230 + }, + { + "epoch": 0.5652960628938666, + "grad_norm": 3.6627259254455566, + "learning_rate": 0.00016231463571889104, + "loss": 0.2118, + "step": 36240 + }, + { + "epoch": 0.5654520496661883, + "grad_norm": 0.3443094789981842, + "learning_rate": 0.00016230423660073626, + "loss": 0.2418, + "step": 36250 + }, + { + "epoch": 0.56560803643851, + "grad_norm": 1.7411112785339355, + "learning_rate": 0.00016229383748258148, + "loss": 0.1628, + "step": 36260 + }, + { + "epoch": 0.5657640232108317, + "grad_norm": 1.1321816444396973, + "learning_rate": 0.0001622834383644267, + "loss": 0.2732, + "step": 36270 + }, + { + "epoch": 0.5659200099831534, + "grad_norm": 2.9686615467071533, + "learning_rate": 0.00016227303924627192, + "loss": 0.3036, + "step": 36280 + }, + { + "epoch": 0.5660759967554752, + "grad_norm": 5.03530740737915, + "learning_rate": 0.00016226264012811714, + "loss": 0.2143, + "step": 36290 + }, + { + "epoch": 0.5662319835277968, + "grad_norm": 2.3943281173706055, + "learning_rate": 0.00016225224100996236, + "loss": 0.2629, + "step": 36300 + }, + { + "epoch": 0.5663879703001186, + "grad_norm": 1.812828779220581, + "learning_rate": 0.00016224184189180758, + "loss": 0.3892, + "step": 36310 + }, + { + "epoch": 0.5665439570724402, + "grad_norm": 1.4708483219146729, + "learning_rate": 0.0001622314427736528, + "loss": 0.2118, + "step": 36320 + }, + { + "epoch": 0.566699943844762, + "grad_norm": 8.99713134765625, + "learning_rate": 0.00016222104365549802, + "loss": 0.3185, + "step": 36330 + }, + { + "epoch": 0.5668559306170837, + "grad_norm": 1.7472341060638428, + "learning_rate": 0.00016221064453734324, + "loss": 0.1659, + "step": 36340 + }, + { + "epoch": 0.5670119173894054, + "grad_norm": 0.8059778213500977, + "learning_rate": 0.00016220024541918846, + "loss": 0.0621, + "step": 36350 + }, + { + "epoch": 0.5671679041617271, + "grad_norm": 4.0501861572265625, + "learning_rate": 0.00016218984630103368, + "loss": 0.1399, + "step": 36360 + }, + { + "epoch": 0.5673238909340488, + "grad_norm": 0.014400321058928967, + "learning_rate": 0.0001621794471828789, + "loss": 0.2372, + "step": 36370 + }, + { + "epoch": 0.5674798777063705, + "grad_norm": 0.4061933755874634, + "learning_rate": 0.00016216904806472412, + "loss": 0.3854, + "step": 36380 + }, + { + "epoch": 0.5676358644786922, + "grad_norm": 0.4940955340862274, + "learning_rate": 0.00016215864894656933, + "loss": 0.1689, + "step": 36390 + }, + { + "epoch": 0.567791851251014, + "grad_norm": 0.3220517635345459, + "learning_rate": 0.00016214824982841455, + "loss": 0.1912, + "step": 36400 + }, + { + "epoch": 0.5679478380233356, + "grad_norm": 2.0848920345306396, + "learning_rate": 0.00016213785071025977, + "loss": 0.1639, + "step": 36410 + }, + { + "epoch": 0.5681038247956574, + "grad_norm": 0.04129406809806824, + "learning_rate": 0.000162127451592105, + "loss": 0.1378, + "step": 36420 + }, + { + "epoch": 0.568259811567979, + "grad_norm": 1.9343401193618774, + "learning_rate": 0.0001621170524739502, + "loss": 0.3759, + "step": 36430 + }, + { + "epoch": 0.5684157983403008, + "grad_norm": 1.0625615119934082, + "learning_rate": 0.00016210665335579543, + "loss": 0.2074, + "step": 36440 + }, + { + "epoch": 0.5685717851126224, + "grad_norm": 0.14784491062164307, + "learning_rate": 0.00016209625423764065, + "loss": 0.0981, + "step": 36450 + }, + { + "epoch": 0.5687277718849442, + "grad_norm": 1.8117821216583252, + "learning_rate": 0.00016208585511948587, + "loss": 0.2755, + "step": 36460 + }, + { + "epoch": 0.5688837586572658, + "grad_norm": 0.061010006815195084, + "learning_rate": 0.0001620754560013311, + "loss": 0.1815, + "step": 36470 + }, + { + "epoch": 0.5690397454295876, + "grad_norm": 0.44453224539756775, + "learning_rate": 0.0001620650568831763, + "loss": 0.0891, + "step": 36480 + }, + { + "epoch": 0.5691957322019093, + "grad_norm": 0.6130645871162415, + "learning_rate": 0.00016205465776502153, + "loss": 0.2778, + "step": 36490 + }, + { + "epoch": 0.569351718974231, + "grad_norm": 5.408796787261963, + "learning_rate": 0.00016204425864686678, + "loss": 0.2301, + "step": 36500 + }, + { + "epoch": 0.5695077057465527, + "grad_norm": 1.2095412015914917, + "learning_rate": 0.00016203385952871197, + "loss": 0.1378, + "step": 36510 + }, + { + "epoch": 0.5696636925188744, + "grad_norm": 0.46873530745506287, + "learning_rate": 0.0001620234604105572, + "loss": 0.1431, + "step": 36520 + }, + { + "epoch": 0.5698196792911961, + "grad_norm": 1.028438925743103, + "learning_rate": 0.0001620130612924024, + "loss": 0.1114, + "step": 36530 + }, + { + "epoch": 0.5699756660635178, + "grad_norm": 0.5952587127685547, + "learning_rate": 0.00016200266217424763, + "loss": 0.2156, + "step": 36540 + }, + { + "epoch": 0.5701316528358396, + "grad_norm": 0.5134285688400269, + "learning_rate": 0.00016199226305609285, + "loss": 0.2712, + "step": 36550 + }, + { + "epoch": 0.5702876396081612, + "grad_norm": 0.03440163657069206, + "learning_rate": 0.00016198186393793807, + "loss": 0.312, + "step": 36560 + }, + { + "epoch": 0.570443626380483, + "grad_norm": 0.10363951325416565, + "learning_rate": 0.0001619714648197833, + "loss": 0.2399, + "step": 36570 + }, + { + "epoch": 0.5705996131528046, + "grad_norm": 1.3341947793960571, + "learning_rate": 0.0001619610657016285, + "loss": 0.1397, + "step": 36580 + }, + { + "epoch": 0.5707555999251264, + "grad_norm": 0.0668073296546936, + "learning_rate": 0.00016195066658347373, + "loss": 0.1091, + "step": 36590 + }, + { + "epoch": 0.570911586697448, + "grad_norm": 5.0486674308776855, + "learning_rate": 0.00016194026746531895, + "loss": 0.4893, + "step": 36600 + }, + { + "epoch": 0.5710675734697698, + "grad_norm": 1.2655692100524902, + "learning_rate": 0.00016192986834716417, + "loss": 0.3344, + "step": 36610 + }, + { + "epoch": 0.5712235602420914, + "grad_norm": 0.1458190381526947, + "learning_rate": 0.00016191946922900939, + "loss": 0.247, + "step": 36620 + }, + { + "epoch": 0.5713795470144132, + "grad_norm": 2.5789072513580322, + "learning_rate": 0.0001619090701108546, + "loss": 0.1546, + "step": 36630 + }, + { + "epoch": 0.5715355337867349, + "grad_norm": 1.191821575164795, + "learning_rate": 0.00016189867099269983, + "loss": 0.2215, + "step": 36640 + }, + { + "epoch": 0.5716915205590566, + "grad_norm": 8.163617134094238, + "learning_rate": 0.00016188827187454505, + "loss": 0.1696, + "step": 36650 + }, + { + "epoch": 0.5718475073313783, + "grad_norm": 3.162277936935425, + "learning_rate": 0.00016187787275639027, + "loss": 0.2906, + "step": 36660 + }, + { + "epoch": 0.5720034941037, + "grad_norm": 2.2368602752685547, + "learning_rate": 0.00016186747363823548, + "loss": 0.3806, + "step": 36670 + }, + { + "epoch": 0.5721594808760218, + "grad_norm": 2.0193803310394287, + "learning_rate": 0.0001618570745200807, + "loss": 0.1964, + "step": 36680 + }, + { + "epoch": 0.5723154676483434, + "grad_norm": 0.9131811261177063, + "learning_rate": 0.00016184667540192592, + "loss": 0.1902, + "step": 36690 + }, + { + "epoch": 0.5724714544206652, + "grad_norm": 0.21330870687961578, + "learning_rate": 0.00016183627628377114, + "loss": 0.2907, + "step": 36700 + }, + { + "epoch": 0.5726274411929868, + "grad_norm": 0.7145895957946777, + "learning_rate": 0.00016182587716561636, + "loss": 0.2574, + "step": 36710 + }, + { + "epoch": 0.5727834279653086, + "grad_norm": 1.0699433088302612, + "learning_rate": 0.00016181547804746158, + "loss": 0.1675, + "step": 36720 + }, + { + "epoch": 0.5729394147376302, + "grad_norm": 0.5066679120063782, + "learning_rate": 0.0001618050789293068, + "loss": 0.4487, + "step": 36730 + }, + { + "epoch": 0.573095401509952, + "grad_norm": 0.7482333183288574, + "learning_rate": 0.00016179467981115202, + "loss": 0.4101, + "step": 36740 + }, + { + "epoch": 0.5732513882822736, + "grad_norm": 0.15925532579421997, + "learning_rate": 0.00016178428069299724, + "loss": 0.2825, + "step": 36750 + }, + { + "epoch": 0.5734073750545954, + "grad_norm": 0.95689457654953, + "learning_rate": 0.00016177388157484246, + "loss": 0.2754, + "step": 36760 + }, + { + "epoch": 0.573563361826917, + "grad_norm": 2.15043568611145, + "learning_rate": 0.00016176348245668768, + "loss": 0.2073, + "step": 36770 + }, + { + "epoch": 0.5737193485992388, + "grad_norm": 2.651520252227783, + "learning_rate": 0.0001617530833385329, + "loss": 0.2517, + "step": 36780 + }, + { + "epoch": 0.5738753353715605, + "grad_norm": 2.1156702041625977, + "learning_rate": 0.00016174268422037812, + "loss": 0.3029, + "step": 36790 + }, + { + "epoch": 0.5740313221438822, + "grad_norm": 3.6200709342956543, + "learning_rate": 0.00016173228510222334, + "loss": 0.4174, + "step": 36800 + }, + { + "epoch": 0.574187308916204, + "grad_norm": 1.447937250137329, + "learning_rate": 0.00016172188598406856, + "loss": 0.1763, + "step": 36810 + }, + { + "epoch": 0.5743432956885256, + "grad_norm": 1.4020819664001465, + "learning_rate": 0.00016171148686591378, + "loss": 0.3994, + "step": 36820 + }, + { + "epoch": 0.5744992824608474, + "grad_norm": 1.4079807996749878, + "learning_rate": 0.000161701087747759, + "loss": 0.2051, + "step": 36830 + }, + { + "epoch": 0.574655269233169, + "grad_norm": 0.05268567427992821, + "learning_rate": 0.00016169068862960422, + "loss": 0.1773, + "step": 36840 + }, + { + "epoch": 0.5748112560054908, + "grad_norm": 2.248767614364624, + "learning_rate": 0.00016168028951144944, + "loss": 0.2534, + "step": 36850 + }, + { + "epoch": 0.5749672427778124, + "grad_norm": 0.3726302981376648, + "learning_rate": 0.00016166989039329466, + "loss": 0.328, + "step": 36860 + }, + { + "epoch": 0.5751232295501342, + "grad_norm": 1.0772110223770142, + "learning_rate": 0.00016165949127513988, + "loss": 0.2069, + "step": 36870 + }, + { + "epoch": 0.5752792163224558, + "grad_norm": 0.11339398473501205, + "learning_rate": 0.0001616490921569851, + "loss": 0.1827, + "step": 36880 + }, + { + "epoch": 0.5754352030947776, + "grad_norm": 1.2981103658676147, + "learning_rate": 0.00016163869303883032, + "loss": 0.1378, + "step": 36890 + }, + { + "epoch": 0.5755911898670992, + "grad_norm": 0.3302285075187683, + "learning_rate": 0.00016162829392067554, + "loss": 0.2311, + "step": 36900 + }, + { + "epoch": 0.575747176639421, + "grad_norm": 0.10168436169624329, + "learning_rate": 0.00016161789480252076, + "loss": 0.1678, + "step": 36910 + }, + { + "epoch": 0.5759031634117426, + "grad_norm": 1.6778357028961182, + "learning_rate": 0.00016160749568436598, + "loss": 0.2385, + "step": 36920 + }, + { + "epoch": 0.5760591501840644, + "grad_norm": 8.1710205078125, + "learning_rate": 0.0001615970965662112, + "loss": 0.3287, + "step": 36930 + }, + { + "epoch": 0.5762151369563862, + "grad_norm": 2.6481053829193115, + "learning_rate": 0.00016158669744805641, + "loss": 0.3119, + "step": 36940 + }, + { + "epoch": 0.5763711237287078, + "grad_norm": 0.3120105266571045, + "learning_rate": 0.00016157629832990163, + "loss": 0.1985, + "step": 36950 + }, + { + "epoch": 0.5765271105010296, + "grad_norm": 2.7221410274505615, + "learning_rate": 0.00016156589921174685, + "loss": 0.3697, + "step": 36960 + }, + { + "epoch": 0.5766830972733512, + "grad_norm": 0.34397706389427185, + "learning_rate": 0.00016155550009359207, + "loss": 0.2162, + "step": 36970 + }, + { + "epoch": 0.576839084045673, + "grad_norm": 4.6421003341674805, + "learning_rate": 0.0001615451009754373, + "loss": 0.2892, + "step": 36980 + }, + { + "epoch": 0.5769950708179946, + "grad_norm": 1.5395855903625488, + "learning_rate": 0.0001615347018572825, + "loss": 0.1258, + "step": 36990 + }, + { + "epoch": 0.5771510575903164, + "grad_norm": 0.4586782455444336, + "learning_rate": 0.00016152430273912773, + "loss": 0.1752, + "step": 37000 + }, + { + "epoch": 0.577307044362638, + "grad_norm": 0.4277658760547638, + "learning_rate": 0.00016151390362097295, + "loss": 0.1722, + "step": 37010 + }, + { + "epoch": 0.5774630311349598, + "grad_norm": 1.0282249450683594, + "learning_rate": 0.00016150350450281817, + "loss": 0.1361, + "step": 37020 + }, + { + "epoch": 0.5776190179072814, + "grad_norm": 0.2496921867132187, + "learning_rate": 0.0001614931053846634, + "loss": 0.0954, + "step": 37030 + }, + { + "epoch": 0.5777750046796032, + "grad_norm": 1.7900562286376953, + "learning_rate": 0.0001614827062665086, + "loss": 0.1284, + "step": 37040 + }, + { + "epoch": 0.5779309914519248, + "grad_norm": 0.6629091501235962, + "learning_rate": 0.00016147230714835383, + "loss": 0.2657, + "step": 37050 + }, + { + "epoch": 0.5780869782242466, + "grad_norm": 2.502882242202759, + "learning_rate": 0.00016146190803019905, + "loss": 0.207, + "step": 37060 + }, + { + "epoch": 0.5782429649965682, + "grad_norm": 1.486069917678833, + "learning_rate": 0.00016145150891204427, + "loss": 0.1232, + "step": 37070 + }, + { + "epoch": 0.57839895176889, + "grad_norm": 0.13543163239955902, + "learning_rate": 0.0001614411097938895, + "loss": 0.3245, + "step": 37080 + }, + { + "epoch": 0.5785549385412118, + "grad_norm": 3.9703898429870605, + "learning_rate": 0.0001614307106757347, + "loss": 0.3104, + "step": 37090 + }, + { + "epoch": 0.5787109253135334, + "grad_norm": 3.2236313819885254, + "learning_rate": 0.00016142031155757993, + "loss": 0.1739, + "step": 37100 + }, + { + "epoch": 0.5788669120858552, + "grad_norm": 0.3854866325855255, + "learning_rate": 0.00016140991243942515, + "loss": 0.084, + "step": 37110 + }, + { + "epoch": 0.5790228988581768, + "grad_norm": 0.1981169581413269, + "learning_rate": 0.00016139951332127037, + "loss": 0.2367, + "step": 37120 + }, + { + "epoch": 0.5791788856304986, + "grad_norm": 2.7315785884857178, + "learning_rate": 0.0001613891142031156, + "loss": 0.1816, + "step": 37130 + }, + { + "epoch": 0.5793348724028202, + "grad_norm": 0.48251378536224365, + "learning_rate": 0.0001613787150849608, + "loss": 0.2123, + "step": 37140 + }, + { + "epoch": 0.579490859175142, + "grad_norm": 5.0431413650512695, + "learning_rate": 0.00016136831596680603, + "loss": 0.2489, + "step": 37150 + }, + { + "epoch": 0.5796468459474636, + "grad_norm": 0.81981360912323, + "learning_rate": 0.00016135791684865125, + "loss": 0.5235, + "step": 37160 + }, + { + "epoch": 0.5798028327197854, + "grad_norm": 0.5701965689659119, + "learning_rate": 0.00016134751773049647, + "loss": 0.1709, + "step": 37170 + }, + { + "epoch": 0.579958819492107, + "grad_norm": 1.3905616998672485, + "learning_rate": 0.00016133711861234169, + "loss": 0.2641, + "step": 37180 + }, + { + "epoch": 0.5801148062644288, + "grad_norm": 2.281285047531128, + "learning_rate": 0.0001613267194941869, + "loss": 0.3915, + "step": 37190 + }, + { + "epoch": 0.5802707930367504, + "grad_norm": 1.2038309574127197, + "learning_rate": 0.00016131632037603213, + "loss": 0.0881, + "step": 37200 + }, + { + "epoch": 0.5804267798090722, + "grad_norm": 0.3096400201320648, + "learning_rate": 0.00016130592125787735, + "loss": 0.1341, + "step": 37210 + }, + { + "epoch": 0.5805827665813938, + "grad_norm": 0.10219905525445938, + "learning_rate": 0.00016129552213972256, + "loss": 0.0867, + "step": 37220 + }, + { + "epoch": 0.5807387533537156, + "grad_norm": 0.13869646191596985, + "learning_rate": 0.00016128512302156778, + "loss": 0.1621, + "step": 37230 + }, + { + "epoch": 0.5808947401260373, + "grad_norm": 0.7992938160896301, + "learning_rate": 0.000161274723903413, + "loss": 0.275, + "step": 37240 + }, + { + "epoch": 0.581050726898359, + "grad_norm": 0.2635972201824188, + "learning_rate": 0.00016126432478525822, + "loss": 0.2429, + "step": 37250 + }, + { + "epoch": 0.5812067136706808, + "grad_norm": 1.7211238145828247, + "learning_rate": 0.00016125392566710344, + "loss": 0.1541, + "step": 37260 + }, + { + "epoch": 0.5813627004430024, + "grad_norm": 2.2448506355285645, + "learning_rate": 0.00016124352654894866, + "loss": 0.3689, + "step": 37270 + }, + { + "epoch": 0.5815186872153242, + "grad_norm": 1.238708257675171, + "learning_rate": 0.00016123312743079388, + "loss": 0.3292, + "step": 37280 + }, + { + "epoch": 0.5816746739876458, + "grad_norm": 0.5670503377914429, + "learning_rate": 0.0001612227283126391, + "loss": 0.1151, + "step": 37290 + }, + { + "epoch": 0.5818306607599676, + "grad_norm": 2.287304401397705, + "learning_rate": 0.00016121232919448432, + "loss": 0.3558, + "step": 37300 + }, + { + "epoch": 0.5819866475322892, + "grad_norm": 0.8009629845619202, + "learning_rate": 0.00016120193007632954, + "loss": 0.1861, + "step": 37310 + }, + { + "epoch": 0.582142634304611, + "grad_norm": 1.1901466846466064, + "learning_rate": 0.00016119153095817476, + "loss": 0.089, + "step": 37320 + }, + { + "epoch": 0.5822986210769326, + "grad_norm": 1.7198141813278198, + "learning_rate": 0.00016118113184001998, + "loss": 0.2931, + "step": 37330 + }, + { + "epoch": 0.5824546078492544, + "grad_norm": 0.8832749724388123, + "learning_rate": 0.0001611707327218652, + "loss": 0.2342, + "step": 37340 + }, + { + "epoch": 0.582610594621576, + "grad_norm": 1.0623290538787842, + "learning_rate": 0.00016116033360371042, + "loss": 0.2317, + "step": 37350 + }, + { + "epoch": 0.5827665813938978, + "grad_norm": 0.5173699855804443, + "learning_rate": 0.00016114993448555564, + "loss": 0.1299, + "step": 37360 + }, + { + "epoch": 0.5829225681662195, + "grad_norm": 0.05118720978498459, + "learning_rate": 0.00016113953536740086, + "loss": 0.0786, + "step": 37370 + }, + { + "epoch": 0.5830785549385412, + "grad_norm": 2.0060107707977295, + "learning_rate": 0.00016112913624924608, + "loss": 0.266, + "step": 37380 + }, + { + "epoch": 0.5832345417108629, + "grad_norm": 5.121693134307861, + "learning_rate": 0.0001611187371310913, + "loss": 0.1824, + "step": 37390 + }, + { + "epoch": 0.5833905284831846, + "grad_norm": 0.3568089008331299, + "learning_rate": 0.00016110833801293652, + "loss": 0.3189, + "step": 37400 + }, + { + "epoch": 0.5835465152555064, + "grad_norm": 0.8631492853164673, + "learning_rate": 0.00016109793889478174, + "loss": 0.1724, + "step": 37410 + }, + { + "epoch": 0.583702502027828, + "grad_norm": 1.9214404821395874, + "learning_rate": 0.00016108753977662693, + "loss": 0.2094, + "step": 37420 + }, + { + "epoch": 0.5838584888001498, + "grad_norm": 1.9055646657943726, + "learning_rate": 0.00016107714065847218, + "loss": 0.1258, + "step": 37430 + }, + { + "epoch": 0.5840144755724714, + "grad_norm": 0.6614194512367249, + "learning_rate": 0.00016106674154031737, + "loss": 0.1966, + "step": 37440 + }, + { + "epoch": 0.5841704623447932, + "grad_norm": 0.286883682012558, + "learning_rate": 0.00016105634242216262, + "loss": 0.588, + "step": 37450 + }, + { + "epoch": 0.5843264491171148, + "grad_norm": 0.5599405765533447, + "learning_rate": 0.0001610459433040078, + "loss": 0.3414, + "step": 37460 + }, + { + "epoch": 0.5844824358894366, + "grad_norm": 1.5908915996551514, + "learning_rate": 0.00016103554418585306, + "loss": 0.1264, + "step": 37470 + }, + { + "epoch": 0.5846384226617583, + "grad_norm": 1.0729930400848389, + "learning_rate": 0.00016102514506769825, + "loss": 0.4815, + "step": 37480 + }, + { + "epoch": 0.58479440943408, + "grad_norm": 0.8912356495857239, + "learning_rate": 0.0001610147459495435, + "loss": 0.2092, + "step": 37490 + }, + { + "epoch": 0.5849503962064017, + "grad_norm": 0.24116197228431702, + "learning_rate": 0.0001610043468313887, + "loss": 0.1552, + "step": 37500 + }, + { + "epoch": 0.5851063829787234, + "grad_norm": 2.245192527770996, + "learning_rate": 0.00016099394771323393, + "loss": 0.0553, + "step": 37510 + }, + { + "epoch": 0.5852623697510451, + "grad_norm": 0.11538795381784439, + "learning_rate": 0.00016098354859507913, + "loss": 0.0889, + "step": 37520 + }, + { + "epoch": 0.5854183565233668, + "grad_norm": 1.0372689962387085, + "learning_rate": 0.00016097314947692437, + "loss": 0.3909, + "step": 37530 + }, + { + "epoch": 0.5855743432956885, + "grad_norm": 1.1946550607681274, + "learning_rate": 0.00016096275035876957, + "loss": 0.2167, + "step": 37540 + }, + { + "epoch": 0.5857303300680102, + "grad_norm": 1.3474591970443726, + "learning_rate": 0.0001609523512406148, + "loss": 0.1561, + "step": 37550 + }, + { + "epoch": 0.585886316840332, + "grad_norm": 0.008936616592109203, + "learning_rate": 0.00016094195212246, + "loss": 0.3991, + "step": 37560 + }, + { + "epoch": 0.5860423036126536, + "grad_norm": 1.8822320699691772, + "learning_rate": 0.00016093155300430525, + "loss": 0.2438, + "step": 37570 + }, + { + "epoch": 0.5861982903849754, + "grad_norm": 1.1707524061203003, + "learning_rate": 0.00016092115388615044, + "loss": 0.308, + "step": 37580 + }, + { + "epoch": 0.586354277157297, + "grad_norm": 1.679641842842102, + "learning_rate": 0.0001609107547679957, + "loss": 0.159, + "step": 37590 + }, + { + "epoch": 0.5865102639296188, + "grad_norm": 1.6885284185409546, + "learning_rate": 0.00016090035564984088, + "loss": 0.157, + "step": 37600 + }, + { + "epoch": 0.5866662507019405, + "grad_norm": 0.3876996636390686, + "learning_rate": 0.00016088995653168613, + "loss": 0.4067, + "step": 37610 + }, + { + "epoch": 0.5868222374742622, + "grad_norm": 1.387480616569519, + "learning_rate": 0.00016087955741353132, + "loss": 0.1898, + "step": 37620 + }, + { + "epoch": 0.5869782242465839, + "grad_norm": 0.03653861582279205, + "learning_rate": 0.00016086915829537657, + "loss": 0.1652, + "step": 37630 + }, + { + "epoch": 0.5871342110189056, + "grad_norm": 0.6258412003517151, + "learning_rate": 0.00016085875917722176, + "loss": 0.2558, + "step": 37640 + }, + { + "epoch": 0.5872901977912273, + "grad_norm": 0.10519164055585861, + "learning_rate": 0.000160848360059067, + "loss": 0.243, + "step": 37650 + }, + { + "epoch": 0.587446184563549, + "grad_norm": 0.7370131015777588, + "learning_rate": 0.0001608379609409122, + "loss": 0.2343, + "step": 37660 + }, + { + "epoch": 0.5876021713358707, + "grad_norm": 0.6884517073631287, + "learning_rate": 0.00016082756182275745, + "loss": 0.1847, + "step": 37670 + }, + { + "epoch": 0.5877581581081924, + "grad_norm": 0.6326120495796204, + "learning_rate": 0.00016081716270460264, + "loss": 0.2258, + "step": 37680 + }, + { + "epoch": 0.5879141448805141, + "grad_norm": 7.697889804840088, + "learning_rate": 0.0001608067635864479, + "loss": 0.2068, + "step": 37690 + }, + { + "epoch": 0.5880701316528358, + "grad_norm": 1.8753420114517212, + "learning_rate": 0.00016079636446829308, + "loss": 0.3087, + "step": 37700 + }, + { + "epoch": 0.5882261184251576, + "grad_norm": 0.8591625094413757, + "learning_rate": 0.00016078596535013833, + "loss": 0.1082, + "step": 37710 + }, + { + "epoch": 0.5883821051974792, + "grad_norm": 0.6124423146247864, + "learning_rate": 0.00016077556623198352, + "loss": 0.1536, + "step": 37720 + }, + { + "epoch": 0.588538091969801, + "grad_norm": 0.1400165557861328, + "learning_rate": 0.00016076516711382877, + "loss": 0.145, + "step": 37730 + }, + { + "epoch": 0.5886940787421227, + "grad_norm": 1.9879004955291748, + "learning_rate": 0.00016075476799567396, + "loss": 0.4986, + "step": 37740 + }, + { + "epoch": 0.5888500655144444, + "grad_norm": 0.08687327802181244, + "learning_rate": 0.0001607443688775192, + "loss": 0.3365, + "step": 37750 + }, + { + "epoch": 0.5890060522867661, + "grad_norm": 0.3442278206348419, + "learning_rate": 0.0001607339697593644, + "loss": 0.2274, + "step": 37760 + }, + { + "epoch": 0.5891620390590878, + "grad_norm": 0.06124915927648544, + "learning_rate": 0.00016072357064120965, + "loss": 0.1915, + "step": 37770 + }, + { + "epoch": 0.5893180258314095, + "grad_norm": 1.0996816158294678, + "learning_rate": 0.00016071317152305484, + "loss": 0.1573, + "step": 37780 + }, + { + "epoch": 0.5894740126037312, + "grad_norm": 0.14738436043262482, + "learning_rate": 0.00016070277240490008, + "loss": 0.2524, + "step": 37790 + }, + { + "epoch": 0.5896299993760529, + "grad_norm": 2.3905467987060547, + "learning_rate": 0.00016069237328674528, + "loss": 0.1838, + "step": 37800 + }, + { + "epoch": 0.5897859861483746, + "grad_norm": 2.0800178050994873, + "learning_rate": 0.00016068197416859052, + "loss": 0.3749, + "step": 37810 + }, + { + "epoch": 0.5899419729206963, + "grad_norm": 3.254204511642456, + "learning_rate": 0.00016067157505043574, + "loss": 0.3777, + "step": 37820 + }, + { + "epoch": 0.590097959693018, + "grad_norm": 2.8447892665863037, + "learning_rate": 0.00016066117593228096, + "loss": 0.5338, + "step": 37830 + }, + { + "epoch": 0.5902539464653397, + "grad_norm": 0.624733030796051, + "learning_rate": 0.00016065077681412618, + "loss": 0.2331, + "step": 37840 + }, + { + "epoch": 0.5904099332376614, + "grad_norm": 0.6863507628440857, + "learning_rate": 0.0001606403776959714, + "loss": 0.1747, + "step": 37850 + }, + { + "epoch": 0.5905659200099832, + "grad_norm": 0.6785943508148193, + "learning_rate": 0.00016062997857781662, + "loss": 0.2064, + "step": 37860 + }, + { + "epoch": 0.5907219067823049, + "grad_norm": 0.5739938616752625, + "learning_rate": 0.00016061957945966181, + "loss": 0.2368, + "step": 37870 + }, + { + "epoch": 0.5908778935546266, + "grad_norm": 0.0616387277841568, + "learning_rate": 0.00016060918034150706, + "loss": 0.2224, + "step": 37880 + }, + { + "epoch": 0.5910338803269483, + "grad_norm": 2.745957136154175, + "learning_rate": 0.00016059878122335225, + "loss": 0.3161, + "step": 37890 + }, + { + "epoch": 0.59118986709927, + "grad_norm": 2.52472186088562, + "learning_rate": 0.0001605883821051975, + "loss": 0.1713, + "step": 37900 + }, + { + "epoch": 0.5913458538715917, + "grad_norm": 1.5691877603530884, + "learning_rate": 0.0001605779829870427, + "loss": 0.1402, + "step": 37910 + }, + { + "epoch": 0.5915018406439134, + "grad_norm": 1.848261833190918, + "learning_rate": 0.00016056758386888794, + "loss": 0.2305, + "step": 37920 + }, + { + "epoch": 0.5916578274162351, + "grad_norm": 0.36622482538223267, + "learning_rate": 0.00016055718475073313, + "loss": 0.1834, + "step": 37930 + }, + { + "epoch": 0.5918138141885568, + "grad_norm": 0.21201446652412415, + "learning_rate": 0.00016054678563257838, + "loss": 0.4682, + "step": 37940 + }, + { + "epoch": 0.5919698009608785, + "grad_norm": 0.38921502232551575, + "learning_rate": 0.00016053638651442357, + "loss": 0.2892, + "step": 37950 + }, + { + "epoch": 0.5921257877332002, + "grad_norm": 0.20777840912342072, + "learning_rate": 0.00016052598739626882, + "loss": 0.104, + "step": 37960 + }, + { + "epoch": 0.5922817745055219, + "grad_norm": 0.7090150713920593, + "learning_rate": 0.000160515588278114, + "loss": 0.2592, + "step": 37970 + }, + { + "epoch": 0.5924377612778436, + "grad_norm": 1.9939709901809692, + "learning_rate": 0.00016050518915995926, + "loss": 0.247, + "step": 37980 + }, + { + "epoch": 0.5925937480501653, + "grad_norm": 0.8696603178977966, + "learning_rate": 0.00016049479004180445, + "loss": 0.301, + "step": 37990 + }, + { + "epoch": 0.592749734822487, + "grad_norm": 0.4119095206260681, + "learning_rate": 0.0001604843909236497, + "loss": 0.2011, + "step": 38000 + }, + { + "epoch": 0.5929057215948088, + "grad_norm": 0.060975607484579086, + "learning_rate": 0.0001604739918054949, + "loss": 0.2286, + "step": 38010 + }, + { + "epoch": 0.5930617083671305, + "grad_norm": 0.7039875388145447, + "learning_rate": 0.00016046359268734014, + "loss": 0.0999, + "step": 38020 + }, + { + "epoch": 0.5932176951394522, + "grad_norm": 0.28402045369148254, + "learning_rate": 0.00016045319356918533, + "loss": 0.1835, + "step": 38030 + }, + { + "epoch": 0.5933736819117739, + "grad_norm": 2.217988967895508, + "learning_rate": 0.00016044279445103058, + "loss": 0.0785, + "step": 38040 + }, + { + "epoch": 0.5935296686840956, + "grad_norm": 1.4856069087982178, + "learning_rate": 0.00016043239533287577, + "loss": 0.2528, + "step": 38050 + }, + { + "epoch": 0.5936856554564173, + "grad_norm": 0.9875519871711731, + "learning_rate": 0.00016042199621472101, + "loss": 0.4445, + "step": 38060 + }, + { + "epoch": 0.593841642228739, + "grad_norm": 2.773216724395752, + "learning_rate": 0.0001604115970965662, + "loss": 0.5643, + "step": 38070 + }, + { + "epoch": 0.5939976290010607, + "grad_norm": 1.7584549188613892, + "learning_rate": 0.00016040119797841145, + "loss": 0.1446, + "step": 38080 + }, + { + "epoch": 0.5941536157733824, + "grad_norm": 1.2456908226013184, + "learning_rate": 0.00016039079886025665, + "loss": 0.1896, + "step": 38090 + }, + { + "epoch": 0.5943096025457041, + "grad_norm": 0.021098516881465912, + "learning_rate": 0.0001603803997421019, + "loss": 0.0876, + "step": 38100 + }, + { + "epoch": 0.5944655893180258, + "grad_norm": 1.8131749629974365, + "learning_rate": 0.00016037000062394709, + "loss": 0.1639, + "step": 38110 + }, + { + "epoch": 0.5946215760903475, + "grad_norm": 0.41819462180137634, + "learning_rate": 0.00016035960150579233, + "loss": 0.1311, + "step": 38120 + }, + { + "epoch": 0.5947775628626693, + "grad_norm": 0.6691219806671143, + "learning_rate": 0.00016034920238763753, + "loss": 0.1625, + "step": 38130 + }, + { + "epoch": 0.5949335496349909, + "grad_norm": 0.07854852080345154, + "learning_rate": 0.00016033880326948277, + "loss": 0.3177, + "step": 38140 + }, + { + "epoch": 0.5950895364073127, + "grad_norm": 0.7819689512252808, + "learning_rate": 0.00016032840415132796, + "loss": 0.4225, + "step": 38150 + }, + { + "epoch": 0.5952455231796344, + "grad_norm": 5.3962836265563965, + "learning_rate": 0.0001603180050331732, + "loss": 0.123, + "step": 38160 + }, + { + "epoch": 0.5954015099519561, + "grad_norm": 0.7460303902626038, + "learning_rate": 0.0001603076059150184, + "loss": 0.2182, + "step": 38170 + }, + { + "epoch": 0.5955574967242778, + "grad_norm": 0.26433685421943665, + "learning_rate": 0.00016029720679686365, + "loss": 0.1894, + "step": 38180 + }, + { + "epoch": 0.5957134834965995, + "grad_norm": 0.8701320290565491, + "learning_rate": 0.00016028680767870884, + "loss": 0.2137, + "step": 38190 + }, + { + "epoch": 0.5958694702689212, + "grad_norm": 3.9774930477142334, + "learning_rate": 0.0001602764085605541, + "loss": 0.4637, + "step": 38200 + }, + { + "epoch": 0.5960254570412429, + "grad_norm": 0.5801145434379578, + "learning_rate": 0.00016026600944239928, + "loss": 0.3036, + "step": 38210 + }, + { + "epoch": 0.5961814438135646, + "grad_norm": 1.1890723705291748, + "learning_rate": 0.00016025561032424453, + "loss": 0.0988, + "step": 38220 + }, + { + "epoch": 0.5963374305858863, + "grad_norm": 0.18521073460578918, + "learning_rate": 0.00016024521120608972, + "loss": 0.1529, + "step": 38230 + }, + { + "epoch": 0.596493417358208, + "grad_norm": 2.2602827548980713, + "learning_rate": 0.00016023481208793497, + "loss": 0.2815, + "step": 38240 + }, + { + "epoch": 0.5966494041305297, + "grad_norm": 0.4060989022254944, + "learning_rate": 0.00016022441296978016, + "loss": 0.3466, + "step": 38250 + }, + { + "epoch": 0.5968053909028515, + "grad_norm": 1.4743244647979736, + "learning_rate": 0.0001602140138516254, + "loss": 0.3561, + "step": 38260 + }, + { + "epoch": 0.5969613776751731, + "grad_norm": 5.080365180969238, + "learning_rate": 0.0001602036147334706, + "loss": 0.2914, + "step": 38270 + }, + { + "epoch": 0.5971173644474949, + "grad_norm": 0.16149812936782837, + "learning_rate": 0.00016019321561531585, + "loss": 0.1871, + "step": 38280 + }, + { + "epoch": 0.5972733512198165, + "grad_norm": 1.7535364627838135, + "learning_rate": 0.00016018281649716104, + "loss": 0.4905, + "step": 38290 + }, + { + "epoch": 0.5974293379921383, + "grad_norm": 0.4103987216949463, + "learning_rate": 0.00016017241737900629, + "loss": 0.3813, + "step": 38300 + }, + { + "epoch": 0.59758532476446, + "grad_norm": 0.5315369367599487, + "learning_rate": 0.00016016201826085148, + "loss": 0.3897, + "step": 38310 + }, + { + "epoch": 0.5977413115367817, + "grad_norm": 3.4553463459014893, + "learning_rate": 0.0001601516191426967, + "loss": 0.2047, + "step": 38320 + }, + { + "epoch": 0.5978972983091034, + "grad_norm": 2.225468397140503, + "learning_rate": 0.00016014122002454192, + "loss": 0.4028, + "step": 38330 + }, + { + "epoch": 0.5980532850814251, + "grad_norm": 2.000885486602783, + "learning_rate": 0.00016013082090638714, + "loss": 0.391, + "step": 38340 + }, + { + "epoch": 0.5982092718537468, + "grad_norm": 0.7886672616004944, + "learning_rate": 0.00016012042178823236, + "loss": 0.2994, + "step": 38350 + }, + { + "epoch": 0.5983652586260685, + "grad_norm": 0.5279917120933533, + "learning_rate": 0.00016011002267007758, + "loss": 0.1855, + "step": 38360 + }, + { + "epoch": 0.5985212453983902, + "grad_norm": 0.9789942502975464, + "learning_rate": 0.0001600996235519228, + "loss": 0.2122, + "step": 38370 + }, + { + "epoch": 0.5986772321707119, + "grad_norm": 0.6289750933647156, + "learning_rate": 0.00016008922443376802, + "loss": 0.3737, + "step": 38380 + }, + { + "epoch": 0.5988332189430337, + "grad_norm": 1.3074733018875122, + "learning_rate": 0.00016007882531561324, + "loss": 0.2382, + "step": 38390 + }, + { + "epoch": 0.5989892057153553, + "grad_norm": 1.745284914970398, + "learning_rate": 0.00016006842619745846, + "loss": 0.3167, + "step": 38400 + }, + { + "epoch": 0.5991451924876771, + "grad_norm": 0.1694452464580536, + "learning_rate": 0.00016005802707930368, + "loss": 0.1632, + "step": 38410 + }, + { + "epoch": 0.5993011792599987, + "grad_norm": 1.55560302734375, + "learning_rate": 0.0001600476279611489, + "loss": 0.2371, + "step": 38420 + }, + { + "epoch": 0.5994571660323205, + "grad_norm": 1.044997215270996, + "learning_rate": 0.00016003722884299411, + "loss": 0.0896, + "step": 38430 + }, + { + "epoch": 0.5996131528046421, + "grad_norm": 0.25250110030174255, + "learning_rate": 0.00016002682972483933, + "loss": 0.3096, + "step": 38440 + }, + { + "epoch": 0.5997691395769639, + "grad_norm": 0.1631859689950943, + "learning_rate": 0.00016001643060668455, + "loss": 0.2004, + "step": 38450 + }, + { + "epoch": 0.5999251263492856, + "grad_norm": 0.34967318177223206, + "learning_rate": 0.00016000603148852977, + "loss": 0.1476, + "step": 38460 + }, + { + "epoch": 0.6000811131216073, + "grad_norm": 1.8604000806808472, + "learning_rate": 0.000159995632370375, + "loss": 0.1884, + "step": 38470 + }, + { + "epoch": 0.600237099893929, + "grad_norm": 1.626848578453064, + "learning_rate": 0.0001599852332522202, + "loss": 0.5452, + "step": 38480 + }, + { + "epoch": 0.6003930866662507, + "grad_norm": 0.41935989260673523, + "learning_rate": 0.00015997483413406543, + "loss": 0.3047, + "step": 38490 + }, + { + "epoch": 0.6005490734385724, + "grad_norm": 1.355086088180542, + "learning_rate": 0.00015996443501591065, + "loss": 0.1246, + "step": 38500 + }, + { + "epoch": 0.6007050602108941, + "grad_norm": 1.5382829904556274, + "learning_rate": 0.00015995403589775587, + "loss": 0.2324, + "step": 38510 + }, + { + "epoch": 0.6008610469832159, + "grad_norm": 1.0250797271728516, + "learning_rate": 0.0001599436367796011, + "loss": 0.181, + "step": 38520 + }, + { + "epoch": 0.6010170337555375, + "grad_norm": 0.07823364436626434, + "learning_rate": 0.0001599332376614463, + "loss": 0.1492, + "step": 38530 + }, + { + "epoch": 0.6011730205278593, + "grad_norm": 0.07573071867227554, + "learning_rate": 0.00015992283854329153, + "loss": 0.1256, + "step": 38540 + }, + { + "epoch": 0.6013290073001809, + "grad_norm": 0.5608569383621216, + "learning_rate": 0.00015991243942513675, + "loss": 0.378, + "step": 38550 + }, + { + "epoch": 0.6014849940725027, + "grad_norm": 0.7001075744628906, + "learning_rate": 0.00015990204030698197, + "loss": 0.2518, + "step": 38560 + }, + { + "epoch": 0.6016409808448243, + "grad_norm": 4.868145942687988, + "learning_rate": 0.0001598916411888272, + "loss": 0.1857, + "step": 38570 + }, + { + "epoch": 0.6017969676171461, + "grad_norm": 0.05653705075383186, + "learning_rate": 0.0001598812420706724, + "loss": 0.1918, + "step": 38580 + }, + { + "epoch": 0.6019529543894677, + "grad_norm": 0.8556832671165466, + "learning_rate": 0.00015987084295251763, + "loss": 0.1926, + "step": 38590 + }, + { + "epoch": 0.6021089411617895, + "grad_norm": 1.1988335847854614, + "learning_rate": 0.00015986044383436285, + "loss": 0.3884, + "step": 38600 + }, + { + "epoch": 0.6022649279341112, + "grad_norm": 1.5978832244873047, + "learning_rate": 0.00015985004471620807, + "loss": 0.1986, + "step": 38610 + }, + { + "epoch": 0.6024209147064329, + "grad_norm": 1.084174633026123, + "learning_rate": 0.0001598396455980533, + "loss": 0.4602, + "step": 38620 + }, + { + "epoch": 0.6025769014787546, + "grad_norm": 1.5116629600524902, + "learning_rate": 0.0001598292464798985, + "loss": 0.292, + "step": 38630 + }, + { + "epoch": 0.6027328882510763, + "grad_norm": 0.23945322632789612, + "learning_rate": 0.00015981884736174373, + "loss": 0.1904, + "step": 38640 + }, + { + "epoch": 0.602888875023398, + "grad_norm": 0.348156601190567, + "learning_rate": 0.00015980844824358895, + "loss": 0.1442, + "step": 38650 + }, + { + "epoch": 0.6030448617957197, + "grad_norm": 1.876736044883728, + "learning_rate": 0.00015979804912543417, + "loss": 0.248, + "step": 38660 + }, + { + "epoch": 0.6032008485680415, + "grad_norm": 0.2564373016357422, + "learning_rate": 0.00015978765000727939, + "loss": 0.2429, + "step": 38670 + }, + { + "epoch": 0.6033568353403631, + "grad_norm": 1.824837327003479, + "learning_rate": 0.0001597772508891246, + "loss": 0.2593, + "step": 38680 + }, + { + "epoch": 0.6035128221126849, + "grad_norm": 1.6967028379440308, + "learning_rate": 0.00015976685177096985, + "loss": 0.3918, + "step": 38690 + }, + { + "epoch": 0.6036688088850065, + "grad_norm": 1.6629718542099, + "learning_rate": 0.00015975645265281504, + "loss": 0.2529, + "step": 38700 + }, + { + "epoch": 0.6038247956573283, + "grad_norm": 0.19047370553016663, + "learning_rate": 0.0001597460535346603, + "loss": 0.4545, + "step": 38710 + }, + { + "epoch": 0.6039807824296499, + "grad_norm": 0.6910511255264282, + "learning_rate": 0.00015973565441650548, + "loss": 0.2226, + "step": 38720 + }, + { + "epoch": 0.6041367692019717, + "grad_norm": 1.9387328624725342, + "learning_rate": 0.00015972525529835073, + "loss": 0.5561, + "step": 38730 + }, + { + "epoch": 0.6042927559742933, + "grad_norm": 1.4491708278656006, + "learning_rate": 0.00015971485618019592, + "loss": 0.2987, + "step": 38740 + }, + { + "epoch": 0.6044487427466151, + "grad_norm": 0.32534459233283997, + "learning_rate": 0.00015970445706204117, + "loss": 0.1683, + "step": 38750 + }, + { + "epoch": 0.6046047295189368, + "grad_norm": 0.17335011065006256, + "learning_rate": 0.00015969405794388636, + "loss": 0.1554, + "step": 38760 + }, + { + "epoch": 0.6047607162912585, + "grad_norm": 0.39481663703918457, + "learning_rate": 0.00015968365882573158, + "loss": 0.186, + "step": 38770 + }, + { + "epoch": 0.6049167030635803, + "grad_norm": 0.10422719269990921, + "learning_rate": 0.0001596732597075768, + "loss": 0.1599, + "step": 38780 + }, + { + "epoch": 0.6050726898359019, + "grad_norm": 2.288064479827881, + "learning_rate": 0.00015966286058942202, + "loss": 0.1944, + "step": 38790 + }, + { + "epoch": 0.6052286766082237, + "grad_norm": 1.0062401294708252, + "learning_rate": 0.00015965246147126724, + "loss": 0.176, + "step": 38800 + }, + { + "epoch": 0.6053846633805453, + "grad_norm": 0.21470843255519867, + "learning_rate": 0.00015964206235311246, + "loss": 0.183, + "step": 38810 + }, + { + "epoch": 0.6055406501528671, + "grad_norm": 2.1730270385742188, + "learning_rate": 0.00015963166323495768, + "loss": 0.1788, + "step": 38820 + }, + { + "epoch": 0.6056966369251887, + "grad_norm": 0.6188749074935913, + "learning_rate": 0.0001596212641168029, + "loss": 0.2746, + "step": 38830 + }, + { + "epoch": 0.6058526236975105, + "grad_norm": 1.283058524131775, + "learning_rate": 0.00015961086499864812, + "loss": 0.1066, + "step": 38840 + }, + { + "epoch": 0.6060086104698321, + "grad_norm": 1.1469789743423462, + "learning_rate": 0.00015960046588049334, + "loss": 0.0692, + "step": 38850 + }, + { + "epoch": 0.6061645972421539, + "grad_norm": 1.9450418949127197, + "learning_rate": 0.00015959006676233856, + "loss": 0.2154, + "step": 38860 + }, + { + "epoch": 0.6063205840144755, + "grad_norm": 7.193401336669922, + "learning_rate": 0.00015957966764418378, + "loss": 0.218, + "step": 38870 + }, + { + "epoch": 0.6064765707867973, + "grad_norm": 3.5315003395080566, + "learning_rate": 0.000159569268526029, + "loss": 0.312, + "step": 38880 + }, + { + "epoch": 0.6066325575591189, + "grad_norm": 1.2873650789260864, + "learning_rate": 0.00015955886940787422, + "loss": 0.2587, + "step": 38890 + }, + { + "epoch": 0.6067885443314407, + "grad_norm": 1.1193941831588745, + "learning_rate": 0.00015954847028971944, + "loss": 0.2358, + "step": 38900 + }, + { + "epoch": 0.6069445311037625, + "grad_norm": 1.0471165180206299, + "learning_rate": 0.00015953807117156466, + "loss": 0.2639, + "step": 38910 + }, + { + "epoch": 0.6071005178760841, + "grad_norm": 1.7760034799575806, + "learning_rate": 0.00015952767205340988, + "loss": 0.248, + "step": 38920 + }, + { + "epoch": 0.6072565046484059, + "grad_norm": 0.050742994993925095, + "learning_rate": 0.0001595172729352551, + "loss": 0.1287, + "step": 38930 + }, + { + "epoch": 0.6074124914207275, + "grad_norm": 0.03016967698931694, + "learning_rate": 0.00015950687381710032, + "loss": 0.1843, + "step": 38940 + }, + { + "epoch": 0.6075684781930493, + "grad_norm": 0.6029847860336304, + "learning_rate": 0.00015949647469894554, + "loss": 0.2294, + "step": 38950 + }, + { + "epoch": 0.6077244649653709, + "grad_norm": 0.2955610454082489, + "learning_rate": 0.00015948607558079076, + "loss": 0.3336, + "step": 38960 + }, + { + "epoch": 0.6078804517376927, + "grad_norm": 1.9104161262512207, + "learning_rate": 0.00015947567646263597, + "loss": 0.189, + "step": 38970 + }, + { + "epoch": 0.6080364385100143, + "grad_norm": 0.09197711199522018, + "learning_rate": 0.0001594652773444812, + "loss": 0.1612, + "step": 38980 + }, + { + "epoch": 0.6081924252823361, + "grad_norm": 1.4233628511428833, + "learning_rate": 0.00015945487822632641, + "loss": 0.1521, + "step": 38990 + }, + { + "epoch": 0.6083484120546577, + "grad_norm": 9.787493705749512, + "learning_rate": 0.00015944447910817163, + "loss": 0.135, + "step": 39000 + }, + { + "epoch": 0.6085043988269795, + "grad_norm": 1.3848742246627808, + "learning_rate": 0.00015943407999001685, + "loss": 0.2268, + "step": 39010 + }, + { + "epoch": 0.6086603855993011, + "grad_norm": 0.1524626612663269, + "learning_rate": 0.00015942368087186207, + "loss": 0.3154, + "step": 39020 + }, + { + "epoch": 0.6088163723716229, + "grad_norm": 2.1885435581207275, + "learning_rate": 0.0001594132817537073, + "loss": 0.3174, + "step": 39030 + }, + { + "epoch": 0.6089723591439445, + "grad_norm": 1.4719898700714111, + "learning_rate": 0.0001594028826355525, + "loss": 0.2959, + "step": 39040 + }, + { + "epoch": 0.6091283459162663, + "grad_norm": 0.7844040989875793, + "learning_rate": 0.00015939248351739773, + "loss": 0.3021, + "step": 39050 + }, + { + "epoch": 0.6092843326885881, + "grad_norm": 0.17660890519618988, + "learning_rate": 0.00015938208439924295, + "loss": 0.3608, + "step": 39060 + }, + { + "epoch": 0.6094403194609097, + "grad_norm": 0.24621741473674774, + "learning_rate": 0.00015937168528108817, + "loss": 0.1461, + "step": 39070 + }, + { + "epoch": 0.6095963062332315, + "grad_norm": 0.7903050184249878, + "learning_rate": 0.0001593612861629334, + "loss": 0.1195, + "step": 39080 + }, + { + "epoch": 0.6097522930055531, + "grad_norm": 0.04043950140476227, + "learning_rate": 0.0001593508870447786, + "loss": 0.1938, + "step": 39090 + }, + { + "epoch": 0.6099082797778749, + "grad_norm": 0.7602111101150513, + "learning_rate": 0.00015934048792662383, + "loss": 0.1732, + "step": 39100 + }, + { + "epoch": 0.6100642665501965, + "grad_norm": 0.7902756333351135, + "learning_rate": 0.00015933008880846905, + "loss": 0.2351, + "step": 39110 + }, + { + "epoch": 0.6102202533225183, + "grad_norm": 1.7035608291625977, + "learning_rate": 0.00015931968969031427, + "loss": 0.2409, + "step": 39120 + }, + { + "epoch": 0.6103762400948399, + "grad_norm": 0.7998217940330505, + "learning_rate": 0.0001593092905721595, + "loss": 0.3862, + "step": 39130 + }, + { + "epoch": 0.6105322268671617, + "grad_norm": 0.3530103862285614, + "learning_rate": 0.0001592988914540047, + "loss": 0.2176, + "step": 39140 + }, + { + "epoch": 0.6106882136394833, + "grad_norm": 0.9218409657478333, + "learning_rate": 0.00015928849233584993, + "loss": 0.1362, + "step": 39150 + }, + { + "epoch": 0.6108442004118051, + "grad_norm": 0.643829345703125, + "learning_rate": 0.00015927809321769515, + "loss": 0.2309, + "step": 39160 + }, + { + "epoch": 0.6110001871841267, + "grad_norm": 0.1810176521539688, + "learning_rate": 0.00015926769409954037, + "loss": 0.2461, + "step": 39170 + }, + { + "epoch": 0.6111561739564485, + "grad_norm": 1.3435649871826172, + "learning_rate": 0.0001592572949813856, + "loss": 0.2937, + "step": 39180 + }, + { + "epoch": 0.6113121607287701, + "grad_norm": 1.1048580408096313, + "learning_rate": 0.0001592468958632308, + "loss": 0.2983, + "step": 39190 + }, + { + "epoch": 0.6114681475010919, + "grad_norm": 0.09463644027709961, + "learning_rate": 0.00015923649674507603, + "loss": 0.2318, + "step": 39200 + }, + { + "epoch": 0.6116241342734137, + "grad_norm": 0.9329594969749451, + "learning_rate": 0.00015922609762692125, + "loss": 0.3095, + "step": 39210 + }, + { + "epoch": 0.6117801210457353, + "grad_norm": 1.8576663732528687, + "learning_rate": 0.00015921569850876647, + "loss": 0.2336, + "step": 39220 + }, + { + "epoch": 0.6119361078180571, + "grad_norm": 1.6159908771514893, + "learning_rate": 0.00015920529939061169, + "loss": 0.1745, + "step": 39230 + }, + { + "epoch": 0.6120920945903787, + "grad_norm": 1.289712905883789, + "learning_rate": 0.0001591949002724569, + "loss": 0.2031, + "step": 39240 + }, + { + "epoch": 0.6122480813627005, + "grad_norm": 0.4183761179447174, + "learning_rate": 0.00015918450115430212, + "loss": 0.2911, + "step": 39250 + }, + { + "epoch": 0.6124040681350221, + "grad_norm": 2.125042200088501, + "learning_rate": 0.00015917410203614734, + "loss": 0.1086, + "step": 39260 + }, + { + "epoch": 0.6125600549073439, + "grad_norm": 3.527683973312378, + "learning_rate": 0.00015916370291799256, + "loss": 0.3356, + "step": 39270 + }, + { + "epoch": 0.6127160416796655, + "grad_norm": 1.1525741815567017, + "learning_rate": 0.00015915330379983778, + "loss": 0.222, + "step": 39280 + }, + { + "epoch": 0.6128720284519873, + "grad_norm": 0.5210650563240051, + "learning_rate": 0.000159142904681683, + "loss": 0.1563, + "step": 39290 + }, + { + "epoch": 0.6130280152243089, + "grad_norm": 0.2881133258342743, + "learning_rate": 0.00015913250556352822, + "loss": 0.346, + "step": 39300 + }, + { + "epoch": 0.6131840019966307, + "grad_norm": 1.7034498453140259, + "learning_rate": 0.00015912210644537344, + "loss": 0.1822, + "step": 39310 + }, + { + "epoch": 0.6133399887689523, + "grad_norm": 2.6305196285247803, + "learning_rate": 0.00015911170732721866, + "loss": 0.3373, + "step": 39320 + }, + { + "epoch": 0.6134959755412741, + "grad_norm": 0.4541589915752411, + "learning_rate": 0.00015910130820906388, + "loss": 0.2146, + "step": 39330 + }, + { + "epoch": 0.6136519623135958, + "grad_norm": 0.03301222622394562, + "learning_rate": 0.0001590909090909091, + "loss": 0.1787, + "step": 39340 + }, + { + "epoch": 0.6138079490859175, + "grad_norm": 0.7106437087059021, + "learning_rate": 0.00015908050997275432, + "loss": 0.2315, + "step": 39350 + }, + { + "epoch": 0.6139639358582393, + "grad_norm": 0.381320059299469, + "learning_rate": 0.00015907011085459954, + "loss": 0.1917, + "step": 39360 + }, + { + "epoch": 0.6141199226305609, + "grad_norm": 0.5418739914894104, + "learning_rate": 0.00015905971173644476, + "loss": 0.1571, + "step": 39370 + }, + { + "epoch": 0.6142759094028827, + "grad_norm": 0.3135250210762024, + "learning_rate": 0.00015904931261828998, + "loss": 0.0858, + "step": 39380 + }, + { + "epoch": 0.6144318961752043, + "grad_norm": 1.6296218633651733, + "learning_rate": 0.0001590389135001352, + "loss": 0.2297, + "step": 39390 + }, + { + "epoch": 0.6145878829475261, + "grad_norm": 0.5768391489982605, + "learning_rate": 0.00015902851438198042, + "loss": 0.1699, + "step": 39400 + }, + { + "epoch": 0.6147438697198477, + "grad_norm": 25.794370651245117, + "learning_rate": 0.00015901811526382564, + "loss": 0.1903, + "step": 39410 + }, + { + "epoch": 0.6148998564921695, + "grad_norm": 0.46698322892189026, + "learning_rate": 0.00015900771614567086, + "loss": 0.1415, + "step": 39420 + }, + { + "epoch": 0.6150558432644911, + "grad_norm": 1.3042336702346802, + "learning_rate": 0.00015899731702751608, + "loss": 0.1604, + "step": 39430 + }, + { + "epoch": 0.6152118300368129, + "grad_norm": 1.2311556339263916, + "learning_rate": 0.0001589869179093613, + "loss": 0.2555, + "step": 39440 + }, + { + "epoch": 0.6153678168091345, + "grad_norm": 0.6290935277938843, + "learning_rate": 0.00015897651879120652, + "loss": 0.1723, + "step": 39450 + }, + { + "epoch": 0.6155238035814563, + "grad_norm": 1.4786419868469238, + "learning_rate": 0.00015896611967305174, + "loss": 0.3717, + "step": 39460 + }, + { + "epoch": 0.615679790353778, + "grad_norm": 0.7589457631111145, + "learning_rate": 0.00015895572055489696, + "loss": 0.1912, + "step": 39470 + }, + { + "epoch": 0.6158357771260997, + "grad_norm": 0.10508652776479721, + "learning_rate": 0.00015894532143674218, + "loss": 0.1353, + "step": 39480 + }, + { + "epoch": 0.6159917638984214, + "grad_norm": 0.07394446432590485, + "learning_rate": 0.0001589349223185874, + "loss": 0.1195, + "step": 39490 + }, + { + "epoch": 0.6161477506707431, + "grad_norm": 0.11420007050037384, + "learning_rate": 0.00015892452320043262, + "loss": 0.0976, + "step": 39500 + }, + { + "epoch": 0.6163037374430649, + "grad_norm": 1.2442359924316406, + "learning_rate": 0.00015891412408227784, + "loss": 0.1084, + "step": 39510 + }, + { + "epoch": 0.6164597242153865, + "grad_norm": 0.4080588221549988, + "learning_rate": 0.00015890372496412306, + "loss": 0.2191, + "step": 39520 + }, + { + "epoch": 0.6166157109877083, + "grad_norm": 0.25214430689811707, + "learning_rate": 0.00015889332584596827, + "loss": 0.5548, + "step": 39530 + }, + { + "epoch": 0.6167716977600299, + "grad_norm": 0.41108590364456177, + "learning_rate": 0.0001588829267278135, + "loss": 0.4213, + "step": 39540 + }, + { + "epoch": 0.6169276845323517, + "grad_norm": 3.3010945320129395, + "learning_rate": 0.00015887252760965871, + "loss": 0.3946, + "step": 39550 + }, + { + "epoch": 0.6170836713046733, + "grad_norm": 0.25975245237350464, + "learning_rate": 0.00015886212849150393, + "loss": 0.3045, + "step": 39560 + }, + { + "epoch": 0.6172396580769951, + "grad_norm": 1.255585789680481, + "learning_rate": 0.00015885172937334915, + "loss": 0.2249, + "step": 39570 + }, + { + "epoch": 0.6173956448493167, + "grad_norm": 1.9135524034500122, + "learning_rate": 0.00015884133025519437, + "loss": 0.2401, + "step": 39580 + }, + { + "epoch": 0.6175516316216385, + "grad_norm": 2.1866586208343506, + "learning_rate": 0.0001588309311370396, + "loss": 0.1726, + "step": 39590 + }, + { + "epoch": 0.6177076183939602, + "grad_norm": 2.487704277038574, + "learning_rate": 0.0001588205320188848, + "loss": 0.2728, + "step": 39600 + }, + { + "epoch": 0.6178636051662819, + "grad_norm": 0.7800028920173645, + "learning_rate": 0.00015881013290073003, + "loss": 0.1522, + "step": 39610 + }, + { + "epoch": 0.6180195919386036, + "grad_norm": 3.468012571334839, + "learning_rate": 0.00015879973378257525, + "loss": 0.1984, + "step": 39620 + }, + { + "epoch": 0.6181755787109253, + "grad_norm": 1.5809861421585083, + "learning_rate": 0.00015878933466442047, + "loss": 0.2632, + "step": 39630 + }, + { + "epoch": 0.618331565483247, + "grad_norm": 0.5371276140213013, + "learning_rate": 0.0001587789355462657, + "loss": 0.1915, + "step": 39640 + }, + { + "epoch": 0.6184875522555687, + "grad_norm": 4.711477279663086, + "learning_rate": 0.0001587685364281109, + "loss": 0.5271, + "step": 39650 + }, + { + "epoch": 0.6186435390278905, + "grad_norm": 1.5581507682800293, + "learning_rate": 0.00015875813730995613, + "loss": 0.2482, + "step": 39660 + }, + { + "epoch": 0.6187995258002121, + "grad_norm": 1.7416914701461792, + "learning_rate": 0.00015874773819180132, + "loss": 0.2018, + "step": 39670 + }, + { + "epoch": 0.6189555125725339, + "grad_norm": 0.6761413216590881, + "learning_rate": 0.00015873733907364657, + "loss": 0.1048, + "step": 39680 + }, + { + "epoch": 0.6191114993448555, + "grad_norm": 1.2344286441802979, + "learning_rate": 0.00015872693995549176, + "loss": 0.1492, + "step": 39690 + }, + { + "epoch": 0.6192674861171773, + "grad_norm": 4.924708843231201, + "learning_rate": 0.000158716540837337, + "loss": 0.2234, + "step": 39700 + }, + { + "epoch": 0.619423472889499, + "grad_norm": 2.3538808822631836, + "learning_rate": 0.0001587061417191822, + "loss": 0.3405, + "step": 39710 + }, + { + "epoch": 0.6195794596618207, + "grad_norm": 0.24888922274112701, + "learning_rate": 0.00015869574260102745, + "loss": 0.1021, + "step": 39720 + }, + { + "epoch": 0.6197354464341424, + "grad_norm": 1.2719457149505615, + "learning_rate": 0.00015868534348287264, + "loss": 0.2832, + "step": 39730 + }, + { + "epoch": 0.6198914332064641, + "grad_norm": 2.80794358253479, + "learning_rate": 0.0001586749443647179, + "loss": 0.2295, + "step": 39740 + }, + { + "epoch": 0.6200474199787858, + "grad_norm": 0.09828778356313705, + "learning_rate": 0.00015866454524656308, + "loss": 0.3574, + "step": 39750 + }, + { + "epoch": 0.6202034067511075, + "grad_norm": 1.0614128112792969, + "learning_rate": 0.00015865414612840833, + "loss": 0.1886, + "step": 39760 + }, + { + "epoch": 0.6203593935234292, + "grad_norm": 0.6939805150032043, + "learning_rate": 0.00015864374701025352, + "loss": 0.3412, + "step": 39770 + }, + { + "epoch": 0.6205153802957509, + "grad_norm": 0.06902116537094116, + "learning_rate": 0.00015863334789209877, + "loss": 0.1647, + "step": 39780 + }, + { + "epoch": 0.6206713670680726, + "grad_norm": 2.05729079246521, + "learning_rate": 0.00015862294877394396, + "loss": 0.168, + "step": 39790 + }, + { + "epoch": 0.6208273538403943, + "grad_norm": 4.69445276260376, + "learning_rate": 0.0001586125496557892, + "loss": 0.1701, + "step": 39800 + }, + { + "epoch": 0.6209833406127161, + "grad_norm": 0.6905967593193054, + "learning_rate": 0.0001586021505376344, + "loss": 0.2823, + "step": 39810 + }, + { + "epoch": 0.6211393273850377, + "grad_norm": 3.0361857414245605, + "learning_rate": 0.00015859175141947964, + "loss": 0.1839, + "step": 39820 + }, + { + "epoch": 0.6212953141573595, + "grad_norm": 0.08303021639585495, + "learning_rate": 0.00015858135230132484, + "loss": 0.2183, + "step": 39830 + }, + { + "epoch": 0.6214513009296811, + "grad_norm": 3.230077028274536, + "learning_rate": 0.00015857095318317008, + "loss": 0.1727, + "step": 39840 + }, + { + "epoch": 0.6216072877020029, + "grad_norm": 1.2884387969970703, + "learning_rate": 0.00015856055406501528, + "loss": 0.1081, + "step": 39850 + }, + { + "epoch": 0.6217632744743246, + "grad_norm": 2.912475109100342, + "learning_rate": 0.00015855015494686052, + "loss": 0.2695, + "step": 39860 + }, + { + "epoch": 0.6219192612466463, + "grad_norm": 0.14122223854064941, + "learning_rate": 0.00015853975582870572, + "loss": 0.3006, + "step": 39870 + }, + { + "epoch": 0.622075248018968, + "grad_norm": 0.8160560131072998, + "learning_rate": 0.00015852935671055096, + "loss": 0.1204, + "step": 39880 + }, + { + "epoch": 0.6222312347912897, + "grad_norm": 1.6930943727493286, + "learning_rate": 0.00015851895759239615, + "loss": 0.1809, + "step": 39890 + }, + { + "epoch": 0.6223872215636114, + "grad_norm": 0.008987984620034695, + "learning_rate": 0.0001585085584742414, + "loss": 0.125, + "step": 39900 + }, + { + "epoch": 0.6225432083359331, + "grad_norm": 2.6068177223205566, + "learning_rate": 0.0001584981593560866, + "loss": 0.3131, + "step": 39910 + }, + { + "epoch": 0.6226991951082548, + "grad_norm": 0.5021041035652161, + "learning_rate": 0.00015848776023793184, + "loss": 0.2023, + "step": 39920 + }, + { + "epoch": 0.6228551818805765, + "grad_norm": 1.8558531999588013, + "learning_rate": 0.00015847736111977703, + "loss": 0.3135, + "step": 39930 + }, + { + "epoch": 0.6230111686528982, + "grad_norm": 0.019155239686369896, + "learning_rate": 0.00015846696200162228, + "loss": 0.2282, + "step": 39940 + }, + { + "epoch": 0.6231671554252199, + "grad_norm": 0.47993865609169006, + "learning_rate": 0.00015845656288346747, + "loss": 0.1143, + "step": 39950 + }, + { + "epoch": 0.6233231421975417, + "grad_norm": 2.553943157196045, + "learning_rate": 0.00015844616376531272, + "loss": 0.2061, + "step": 39960 + }, + { + "epoch": 0.6234791289698633, + "grad_norm": 0.22600245475769043, + "learning_rate": 0.0001584357646471579, + "loss": 0.1808, + "step": 39970 + }, + { + "epoch": 0.6236351157421851, + "grad_norm": 0.7153030037879944, + "learning_rate": 0.00015842536552900316, + "loss": 0.2221, + "step": 39980 + }, + { + "epoch": 0.6237911025145068, + "grad_norm": 0.12053301930427551, + "learning_rate": 0.00015841496641084835, + "loss": 0.2349, + "step": 39990 + }, + { + "epoch": 0.6239470892868285, + "grad_norm": 0.44607964158058167, + "learning_rate": 0.0001584045672926936, + "loss": 0.234, + "step": 40000 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.43583156224e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2c9d27434a451bc559b026c74a98854e5516c3b5 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:24936d92dedebf059c086262587b0ac1dc91b9f24fffae51359f133f351ca389 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..cb91ff0fee22412e9d6b7d9319e19f72cd4d3145 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:36448c418f9c92581dc9739096ce446cebc0f560cb945319263461b66ae4c510 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..e17f00863f8db528ae7dd6a4dfac3d8af478d816 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..3de1b89a616a62b83a6c51d8dfee29e5a8fbd45f Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..dfbaa4a3d1328f7dc2667feab08cc5a0d753ace5 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/trainer_state.json @@ -0,0 +1,28384 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.6317464279029138, + "eval_steps": 500, + "global_step": 40500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + }, + { + "epoch": 0.5071129968178698, + "grad_norm": 1.0075199604034424, + "learning_rate": 0.00016619350679062414, + "loss": 0.2569, + "step": 32510 + }, + { + "epoch": 0.5072689835901916, + "grad_norm": 1.3489066362380981, + "learning_rate": 0.0001661831076724694, + "loss": 0.2565, + "step": 32520 + }, + { + "epoch": 0.5074249703625132, + "grad_norm": 0.2534058094024658, + "learning_rate": 0.00016617270855431458, + "loss": 0.2292, + "step": 32530 + }, + { + "epoch": 0.507580957134835, + "grad_norm": 0.31723281741142273, + "learning_rate": 0.00016616230943615983, + "loss": 0.1467, + "step": 32540 + }, + { + "epoch": 0.5077369439071566, + "grad_norm": 0.29659849405288696, + "learning_rate": 0.00016615191031800502, + "loss": 0.2768, + "step": 32550 + }, + { + "epoch": 0.5078929306794784, + "grad_norm": 1.5246435403823853, + "learning_rate": 0.00016614151119985027, + "loss": 0.216, + "step": 32560 + }, + { + "epoch": 0.5080489174518001, + "grad_norm": 0.6026854515075684, + "learning_rate": 0.00016613111208169546, + "loss": 0.2277, + "step": 32570 + }, + { + "epoch": 0.5082049042241218, + "grad_norm": 0.41994890570640564, + "learning_rate": 0.0001661207129635407, + "loss": 0.2433, + "step": 32580 + }, + { + "epoch": 0.5083608909964435, + "grad_norm": 2.084599733352661, + "learning_rate": 0.00016611031384538593, + "loss": 0.4081, + "step": 32590 + }, + { + "epoch": 0.5085168777687652, + "grad_norm": 0.2945229709148407, + "learning_rate": 0.00016609991472723115, + "loss": 0.1329, + "step": 32600 + }, + { + "epoch": 0.508672864541087, + "grad_norm": 1.4981423616409302, + "learning_rate": 0.00016608951560907636, + "loss": 0.2855, + "step": 32610 + }, + { + "epoch": 0.5088288513134086, + "grad_norm": 0.961737871170044, + "learning_rate": 0.00016607911649092158, + "loss": 0.1654, + "step": 32620 + }, + { + "epoch": 0.5089848380857304, + "grad_norm": 1.49734628200531, + "learning_rate": 0.0001660687173727668, + "loss": 0.2081, + "step": 32630 + }, + { + "epoch": 0.509140824858052, + "grad_norm": 2.538135051727295, + "learning_rate": 0.00016605831825461202, + "loss": 0.3023, + "step": 32640 + }, + { + "epoch": 0.5092968116303738, + "grad_norm": 0.2981272041797638, + "learning_rate": 0.00016604791913645724, + "loss": 0.214, + "step": 32650 + }, + { + "epoch": 0.5094527984026954, + "grad_norm": 1.8863650560379028, + "learning_rate": 0.00016603752001830246, + "loss": 0.3363, + "step": 32660 + }, + { + "epoch": 0.5096087851750172, + "grad_norm": 0.022720983251929283, + "learning_rate": 0.00016602712090014768, + "loss": 0.2463, + "step": 32670 + }, + { + "epoch": 0.5097647719473388, + "grad_norm": 1.154977798461914, + "learning_rate": 0.0001660167217819929, + "loss": 0.4484, + "step": 32680 + }, + { + "epoch": 0.5099207587196606, + "grad_norm": 0.9470499157905579, + "learning_rate": 0.00016600632266383812, + "loss": 0.1196, + "step": 32690 + }, + { + "epoch": 0.5100767454919822, + "grad_norm": 0.19955401122570038, + "learning_rate": 0.00016599592354568334, + "loss": 0.1762, + "step": 32700 + }, + { + "epoch": 0.510232732264304, + "grad_norm": 0.4098282754421234, + "learning_rate": 0.00016598552442752856, + "loss": 0.2739, + "step": 32710 + }, + { + "epoch": 0.5103887190366257, + "grad_norm": 0.21697258949279785, + "learning_rate": 0.00016597512530937378, + "loss": 0.2555, + "step": 32720 + }, + { + "epoch": 0.5105447058089474, + "grad_norm": 4.894479751586914, + "learning_rate": 0.000165964726191219, + "loss": 0.3137, + "step": 32730 + }, + { + "epoch": 0.5107006925812692, + "grad_norm": 0.8396711349487305, + "learning_rate": 0.00016595432707306422, + "loss": 0.3105, + "step": 32740 + }, + { + "epoch": 0.5108566793535908, + "grad_norm": 1.5494674444198608, + "learning_rate": 0.00016594392795490944, + "loss": 0.2271, + "step": 32750 + }, + { + "epoch": 0.5110126661259126, + "grad_norm": 1.3409888744354248, + "learning_rate": 0.00016593352883675466, + "loss": 0.1796, + "step": 32760 + }, + { + "epoch": 0.5111686528982342, + "grad_norm": 1.7590911388397217, + "learning_rate": 0.00016592312971859988, + "loss": 0.2185, + "step": 32770 + }, + { + "epoch": 0.511324639670556, + "grad_norm": 0.7350002527236938, + "learning_rate": 0.0001659127306004451, + "loss": 0.1911, + "step": 32780 + }, + { + "epoch": 0.5114806264428776, + "grad_norm": 0.7176411747932434, + "learning_rate": 0.00016590233148229032, + "loss": 0.6918, + "step": 32790 + }, + { + "epoch": 0.5116366132151994, + "grad_norm": 2.5005321502685547, + "learning_rate": 0.00016589193236413554, + "loss": 0.2329, + "step": 32800 + }, + { + "epoch": 0.511792599987521, + "grad_norm": 1.3916250467300415, + "learning_rate": 0.00016588153324598076, + "loss": 0.2102, + "step": 32810 + }, + { + "epoch": 0.5119485867598428, + "grad_norm": 0.721671998500824, + "learning_rate": 0.00016587113412782598, + "loss": 0.2965, + "step": 32820 + }, + { + "epoch": 0.5121045735321644, + "grad_norm": 0.6120114326477051, + "learning_rate": 0.0001658607350096712, + "loss": 0.2661, + "step": 32830 + }, + { + "epoch": 0.5122605603044862, + "grad_norm": 0.29123347997665405, + "learning_rate": 0.00016585033589151642, + "loss": 0.2455, + "step": 32840 + }, + { + "epoch": 0.5124165470768078, + "grad_norm": 2.7300868034362793, + "learning_rate": 0.00016583993677336164, + "loss": 0.2049, + "step": 32850 + }, + { + "epoch": 0.5125725338491296, + "grad_norm": 1.801928997039795, + "learning_rate": 0.00016582953765520686, + "loss": 0.1725, + "step": 32860 + }, + { + "epoch": 0.5127285206214514, + "grad_norm": 0.6406471729278564, + "learning_rate": 0.00016581913853705208, + "loss": 0.164, + "step": 32870 + }, + { + "epoch": 0.512884507393773, + "grad_norm": 0.9627325534820557, + "learning_rate": 0.0001658087394188973, + "loss": 0.1728, + "step": 32880 + }, + { + "epoch": 0.5130404941660948, + "grad_norm": 1.5446962118148804, + "learning_rate": 0.00016579834030074251, + "loss": 0.2345, + "step": 32890 + }, + { + "epoch": 0.5131964809384164, + "grad_norm": 0.28092706203460693, + "learning_rate": 0.0001657879411825877, + "loss": 0.2124, + "step": 32900 + }, + { + "epoch": 0.5133524677107382, + "grad_norm": 1.7751421928405762, + "learning_rate": 0.00016577754206443295, + "loss": 0.2367, + "step": 32910 + }, + { + "epoch": 0.5135084544830598, + "grad_norm": 1.6964634656906128, + "learning_rate": 0.00016576714294627815, + "loss": 0.2361, + "step": 32920 + }, + { + "epoch": 0.5136644412553816, + "grad_norm": 0.06759823113679886, + "learning_rate": 0.0001657567438281234, + "loss": 0.2329, + "step": 32930 + }, + { + "epoch": 0.5138204280277032, + "grad_norm": 2.2519280910491943, + "learning_rate": 0.00016574634470996859, + "loss": 0.1079, + "step": 32940 + }, + { + "epoch": 0.513976414800025, + "grad_norm": 3.167576551437378, + "learning_rate": 0.00016573594559181383, + "loss": 0.1764, + "step": 32950 + }, + { + "epoch": 0.5141324015723466, + "grad_norm": 1.6734397411346436, + "learning_rate": 0.00016572554647365903, + "loss": 0.2878, + "step": 32960 + }, + { + "epoch": 0.5142883883446684, + "grad_norm": 0.35627374053001404, + "learning_rate": 0.00016571514735550427, + "loss": 0.2337, + "step": 32970 + }, + { + "epoch": 0.51444437511699, + "grad_norm": 1.6321039199829102, + "learning_rate": 0.00016570474823734946, + "loss": 0.2551, + "step": 32980 + }, + { + "epoch": 0.5146003618893118, + "grad_norm": 1.352982997894287, + "learning_rate": 0.0001656943491191947, + "loss": 0.4467, + "step": 32990 + }, + { + "epoch": 0.5147563486616334, + "grad_norm": 0.3558129072189331, + "learning_rate": 0.0001656839500010399, + "loss": 0.1913, + "step": 33000 + }, + { + "epoch": 0.5149123354339552, + "grad_norm": 0.041972286999225616, + "learning_rate": 0.00016567355088288515, + "loss": 0.1704, + "step": 33010 + }, + { + "epoch": 0.515068322206277, + "grad_norm": 1.4168939590454102, + "learning_rate": 0.00016566315176473034, + "loss": 0.4866, + "step": 33020 + }, + { + "epoch": 0.5152243089785986, + "grad_norm": 3.0042929649353027, + "learning_rate": 0.0001656527526465756, + "loss": 0.2473, + "step": 33030 + }, + { + "epoch": 0.5153802957509204, + "grad_norm": 1.027274489402771, + "learning_rate": 0.00016564235352842078, + "loss": 0.436, + "step": 33040 + }, + { + "epoch": 0.515536282523242, + "grad_norm": 1.2330195903778076, + "learning_rate": 0.00016563195441026603, + "loss": 0.4615, + "step": 33050 + }, + { + "epoch": 0.5156922692955638, + "grad_norm": 2.7145066261291504, + "learning_rate": 0.00016562155529211122, + "loss": 0.278, + "step": 33060 + }, + { + "epoch": 0.5158482560678854, + "grad_norm": 1.1273893117904663, + "learning_rate": 0.00016561115617395647, + "loss": 0.3767, + "step": 33070 + }, + { + "epoch": 0.5160042428402072, + "grad_norm": 2.466265916824341, + "learning_rate": 0.00016560075705580166, + "loss": 0.1879, + "step": 33080 + }, + { + "epoch": 0.5161602296125288, + "grad_norm": 0.3759515583515167, + "learning_rate": 0.0001655903579376469, + "loss": 0.4004, + "step": 33090 + }, + { + "epoch": 0.5163162163848506, + "grad_norm": 1.8821336030960083, + "learning_rate": 0.0001655799588194921, + "loss": 0.1622, + "step": 33100 + }, + { + "epoch": 0.5164722031571722, + "grad_norm": 0.8162096738815308, + "learning_rate": 0.00016556955970133735, + "loss": 0.1341, + "step": 33110 + }, + { + "epoch": 0.516628189929494, + "grad_norm": 0.008498801849782467, + "learning_rate": 0.00016555916058318254, + "loss": 0.2546, + "step": 33120 + }, + { + "epoch": 0.5167841767018156, + "grad_norm": 0.17536675930023193, + "learning_rate": 0.00016554876146502779, + "loss": 0.2934, + "step": 33130 + }, + { + "epoch": 0.5169401634741374, + "grad_norm": 3.539898633956909, + "learning_rate": 0.00016553836234687298, + "loss": 0.4959, + "step": 33140 + }, + { + "epoch": 0.517096150246459, + "grad_norm": 2.190887212753296, + "learning_rate": 0.00016552796322871823, + "loss": 0.2914, + "step": 33150 + }, + { + "epoch": 0.5172521370187808, + "grad_norm": 1.6388943195343018, + "learning_rate": 0.00016551756411056342, + "loss": 0.1317, + "step": 33160 + }, + { + "epoch": 0.5174081237911026, + "grad_norm": 0.33063387870788574, + "learning_rate": 0.00016550716499240866, + "loss": 0.4258, + "step": 33170 + }, + { + "epoch": 0.5175641105634242, + "grad_norm": 0.5168288946151733, + "learning_rate": 0.00016549676587425386, + "loss": 0.2548, + "step": 33180 + }, + { + "epoch": 0.517720097335746, + "grad_norm": 3.9570281505584717, + "learning_rate": 0.0001654863667560991, + "loss": 0.3179, + "step": 33190 + }, + { + "epoch": 0.5178760841080676, + "grad_norm": 1.127255916595459, + "learning_rate": 0.0001654759676379443, + "loss": 0.233, + "step": 33200 + }, + { + "epoch": 0.5180320708803894, + "grad_norm": 0.36299410462379456, + "learning_rate": 0.00016546556851978954, + "loss": 0.4614, + "step": 33210 + }, + { + "epoch": 0.518188057652711, + "grad_norm": 2.914668321609497, + "learning_rate": 0.00016545516940163474, + "loss": 0.1603, + "step": 33220 + }, + { + "epoch": 0.5183440444250328, + "grad_norm": 6.991094589233398, + "learning_rate": 0.00016544477028347998, + "loss": 0.2259, + "step": 33230 + }, + { + "epoch": 0.5185000311973544, + "grad_norm": 2.0594241619110107, + "learning_rate": 0.00016543437116532517, + "loss": 0.1885, + "step": 33240 + }, + { + "epoch": 0.5186560179696762, + "grad_norm": 0.5594693422317505, + "learning_rate": 0.00016542397204717042, + "loss": 0.4333, + "step": 33250 + }, + { + "epoch": 0.5188120047419978, + "grad_norm": 4.314854621887207, + "learning_rate": 0.00016541357292901561, + "loss": 0.2103, + "step": 33260 + }, + { + "epoch": 0.5189679915143196, + "grad_norm": 0.5609000325202942, + "learning_rate": 0.00016540317381086086, + "loss": 0.2498, + "step": 33270 + }, + { + "epoch": 0.5191239782866413, + "grad_norm": 3.680088758468628, + "learning_rate": 0.00016539277469270605, + "loss": 0.3465, + "step": 33280 + }, + { + "epoch": 0.519279965058963, + "grad_norm": 1.4993672370910645, + "learning_rate": 0.0001653823755745513, + "loss": 0.3272, + "step": 33290 + }, + { + "epoch": 0.5194359518312847, + "grad_norm": 0.18005084991455078, + "learning_rate": 0.0001653719764563965, + "loss": 0.1706, + "step": 33300 + }, + { + "epoch": 0.5195919386036064, + "grad_norm": 0.5105524659156799, + "learning_rate": 0.00016536157733824174, + "loss": 0.0937, + "step": 33310 + }, + { + "epoch": 0.5197479253759281, + "grad_norm": 1.2806589603424072, + "learning_rate": 0.00016535117822008693, + "loss": 0.1578, + "step": 33320 + }, + { + "epoch": 0.5199039121482498, + "grad_norm": 0.22015990316867828, + "learning_rate": 0.00016534077910193218, + "loss": 0.2081, + "step": 33330 + }, + { + "epoch": 0.5200598989205716, + "grad_norm": 0.34638991951942444, + "learning_rate": 0.00016533037998377737, + "loss": 0.2389, + "step": 33340 + }, + { + "epoch": 0.5202158856928932, + "grad_norm": 0.8709302544593811, + "learning_rate": 0.0001653199808656226, + "loss": 0.2112, + "step": 33350 + }, + { + "epoch": 0.520371872465215, + "grad_norm": 1.2309521436691284, + "learning_rate": 0.0001653095817474678, + "loss": 0.3448, + "step": 33360 + }, + { + "epoch": 0.5205278592375366, + "grad_norm": 1.3155115842819214, + "learning_rate": 0.00016529918262931303, + "loss": 0.1572, + "step": 33370 + }, + { + "epoch": 0.5206838460098584, + "grad_norm": 0.358751118183136, + "learning_rate": 0.00016528878351115825, + "loss": 0.3011, + "step": 33380 + }, + { + "epoch": 0.52083983278218, + "grad_norm": 3.800628662109375, + "learning_rate": 0.00016527838439300347, + "loss": 0.458, + "step": 33390 + }, + { + "epoch": 0.5209958195545018, + "grad_norm": 1.1253124475479126, + "learning_rate": 0.0001652679852748487, + "loss": 0.1238, + "step": 33400 + }, + { + "epoch": 0.5211518063268235, + "grad_norm": 0.46979108452796936, + "learning_rate": 0.0001652575861566939, + "loss": 0.0745, + "step": 33410 + }, + { + "epoch": 0.5213077930991452, + "grad_norm": 7.921492576599121, + "learning_rate": 0.00016524718703853913, + "loss": 0.4004, + "step": 33420 + }, + { + "epoch": 0.5214637798714669, + "grad_norm": 1.643127679824829, + "learning_rate": 0.00016523678792038435, + "loss": 0.2858, + "step": 33430 + }, + { + "epoch": 0.5216197666437886, + "grad_norm": 7.096631050109863, + "learning_rate": 0.0001652263888022296, + "loss": 0.4735, + "step": 33440 + }, + { + "epoch": 0.5217757534161103, + "grad_norm": 0.7280219197273254, + "learning_rate": 0.0001652159896840748, + "loss": 0.1841, + "step": 33450 + }, + { + "epoch": 0.521931740188432, + "grad_norm": 0.9415522813796997, + "learning_rate": 0.00016520559056592003, + "loss": 0.2169, + "step": 33460 + }, + { + "epoch": 0.5220877269607537, + "grad_norm": 2.264862060546875, + "learning_rate": 0.00016519519144776523, + "loss": 0.1902, + "step": 33470 + }, + { + "epoch": 0.5222437137330754, + "grad_norm": 2.4566822052001953, + "learning_rate": 0.00016518479232961047, + "loss": 0.3329, + "step": 33480 + }, + { + "epoch": 0.5223997005053972, + "grad_norm": 0.7330378890037537, + "learning_rate": 0.00016517439321145567, + "loss": 0.2341, + "step": 33490 + }, + { + "epoch": 0.5225556872777188, + "grad_norm": 1.6409542560577393, + "learning_rate": 0.0001651639940933009, + "loss": 0.3077, + "step": 33500 + }, + { + "epoch": 0.5227116740500406, + "grad_norm": 1.0647432804107666, + "learning_rate": 0.0001651535949751461, + "loss": 0.1935, + "step": 33510 + }, + { + "epoch": 0.5228676608223622, + "grad_norm": 0.507420539855957, + "learning_rate": 0.00016514319585699135, + "loss": 0.1314, + "step": 33520 + }, + { + "epoch": 0.523023647594684, + "grad_norm": 1.0064164400100708, + "learning_rate": 0.00016513279673883654, + "loss": 0.344, + "step": 33530 + }, + { + "epoch": 0.5231796343670057, + "grad_norm": 1.1936209201812744, + "learning_rate": 0.0001651223976206818, + "loss": 0.1948, + "step": 33540 + }, + { + "epoch": 0.5233356211393274, + "grad_norm": 0.04992926865816116, + "learning_rate": 0.00016511199850252698, + "loss": 0.1363, + "step": 33550 + }, + { + "epoch": 0.5234916079116491, + "grad_norm": 1.8256275653839111, + "learning_rate": 0.00016510159938437223, + "loss": 0.443, + "step": 33560 + }, + { + "epoch": 0.5236475946839708, + "grad_norm": 4.914908409118652, + "learning_rate": 0.00016509120026621742, + "loss": 0.4654, + "step": 33570 + }, + { + "epoch": 0.5238035814562925, + "grad_norm": 1.1625486612319946, + "learning_rate": 0.00016508080114806267, + "loss": 0.1974, + "step": 33580 + }, + { + "epoch": 0.5239595682286142, + "grad_norm": 6.6192522048950195, + "learning_rate": 0.00016507040202990786, + "loss": 0.2131, + "step": 33590 + }, + { + "epoch": 0.5241155550009359, + "grad_norm": 0.7369065880775452, + "learning_rate": 0.0001650600029117531, + "loss": 0.2371, + "step": 33600 + }, + { + "epoch": 0.5242715417732576, + "grad_norm": 1.5238152742385864, + "learning_rate": 0.0001650496037935983, + "loss": 0.1557, + "step": 33610 + }, + { + "epoch": 0.5244275285455793, + "grad_norm": 1.0418007373809814, + "learning_rate": 0.00016503920467544355, + "loss": 0.1878, + "step": 33620 + }, + { + "epoch": 0.524583515317901, + "grad_norm": 0.790117084980011, + "learning_rate": 0.00016502880555728874, + "loss": 0.2195, + "step": 33630 + }, + { + "epoch": 0.5247395020902228, + "grad_norm": 1.6712257862091064, + "learning_rate": 0.000165018406439134, + "loss": 0.1602, + "step": 33640 + }, + { + "epoch": 0.5248954888625444, + "grad_norm": 0.19236230850219727, + "learning_rate": 0.00016500800732097918, + "loss": 0.2526, + "step": 33650 + }, + { + "epoch": 0.5250514756348662, + "grad_norm": 1.3519701957702637, + "learning_rate": 0.00016499760820282443, + "loss": 0.4686, + "step": 33660 + }, + { + "epoch": 0.5252074624071879, + "grad_norm": 1.694342851638794, + "learning_rate": 0.00016498720908466962, + "loss": 0.1859, + "step": 33670 + }, + { + "epoch": 0.5253634491795096, + "grad_norm": 5.225239276885986, + "learning_rate": 0.00016497680996651487, + "loss": 0.2944, + "step": 33680 + }, + { + "epoch": 0.5255194359518313, + "grad_norm": 2.0208842754364014, + "learning_rate": 0.00016496641084836006, + "loss": 0.2421, + "step": 33690 + }, + { + "epoch": 0.525675422724153, + "grad_norm": 0.7954996228218079, + "learning_rate": 0.0001649560117302053, + "loss": 0.2853, + "step": 33700 + }, + { + "epoch": 0.5258314094964747, + "grad_norm": 2.296086072921753, + "learning_rate": 0.0001649456126120505, + "loss": 0.1333, + "step": 33710 + }, + { + "epoch": 0.5259873962687964, + "grad_norm": 1.1779128313064575, + "learning_rate": 0.00016493521349389574, + "loss": 0.1066, + "step": 33720 + }, + { + "epoch": 0.5261433830411181, + "grad_norm": 0.1756065934896469, + "learning_rate": 0.00016492481437574094, + "loss": 0.1352, + "step": 33730 + }, + { + "epoch": 0.5262993698134398, + "grad_norm": 0.13100725412368774, + "learning_rate": 0.00016491441525758618, + "loss": 0.2399, + "step": 33740 + }, + { + "epoch": 0.5264553565857615, + "grad_norm": 5.532008171081543, + "learning_rate": 0.00016490401613943138, + "loss": 0.2896, + "step": 33750 + }, + { + "epoch": 0.5266113433580832, + "grad_norm": 1.319886565208435, + "learning_rate": 0.00016489361702127662, + "loss": 0.3275, + "step": 33760 + }, + { + "epoch": 0.5267673301304049, + "grad_norm": 1.5550974607467651, + "learning_rate": 0.00016488321790312182, + "loss": 0.2677, + "step": 33770 + }, + { + "epoch": 0.5269233169027266, + "grad_norm": 1.8936737775802612, + "learning_rate": 0.00016487281878496706, + "loss": 0.1955, + "step": 33780 + }, + { + "epoch": 0.5270793036750484, + "grad_norm": 0.3653401732444763, + "learning_rate": 0.00016486241966681226, + "loss": 0.0723, + "step": 33790 + }, + { + "epoch": 0.52723529044737, + "grad_norm": 2.861341714859009, + "learning_rate": 0.00016485202054865747, + "loss": 0.2412, + "step": 33800 + }, + { + "epoch": 0.5273912772196918, + "grad_norm": 1.5291428565979004, + "learning_rate": 0.0001648416214305027, + "loss": 0.0871, + "step": 33810 + }, + { + "epoch": 0.5275472639920135, + "grad_norm": 1.0372581481933594, + "learning_rate": 0.00016483122231234791, + "loss": 0.4705, + "step": 33820 + }, + { + "epoch": 0.5277032507643352, + "grad_norm": 1.1943141222000122, + "learning_rate": 0.00016482082319419313, + "loss": 0.2848, + "step": 33830 + }, + { + "epoch": 0.5278592375366569, + "grad_norm": 1.9008225202560425, + "learning_rate": 0.00016481042407603835, + "loss": 0.139, + "step": 33840 + }, + { + "epoch": 0.5280152243089786, + "grad_norm": 2.132089138031006, + "learning_rate": 0.00016480002495788357, + "loss": 0.2119, + "step": 33850 + }, + { + "epoch": 0.5281712110813003, + "grad_norm": 0.24524426460266113, + "learning_rate": 0.0001647896258397288, + "loss": 0.1643, + "step": 33860 + }, + { + "epoch": 0.528327197853622, + "grad_norm": 1.6469637155532837, + "learning_rate": 0.000164779226721574, + "loss": 0.363, + "step": 33870 + }, + { + "epoch": 0.5284831846259437, + "grad_norm": 0.8767328858375549, + "learning_rate": 0.00016476882760341923, + "loss": 0.2632, + "step": 33880 + }, + { + "epoch": 0.5286391713982654, + "grad_norm": 0.06347586214542389, + "learning_rate": 0.00016475842848526445, + "loss": 0.3204, + "step": 33890 + }, + { + "epoch": 0.5287951581705871, + "grad_norm": 0.09782540798187256, + "learning_rate": 0.00016474802936710967, + "loss": 0.2211, + "step": 33900 + }, + { + "epoch": 0.5289511449429088, + "grad_norm": 3.2998859882354736, + "learning_rate": 0.0001647376302489549, + "loss": 0.265, + "step": 33910 + }, + { + "epoch": 0.5291071317152305, + "grad_norm": 0.43594226241111755, + "learning_rate": 0.0001647272311308001, + "loss": 0.194, + "step": 33920 + }, + { + "epoch": 0.5292631184875523, + "grad_norm": 1.5166605710983276, + "learning_rate": 0.00016471683201264533, + "loss": 0.2675, + "step": 33930 + }, + { + "epoch": 0.529419105259874, + "grad_norm": 0.6056640148162842, + "learning_rate": 0.00016470643289449055, + "loss": 0.0576, + "step": 33940 + }, + { + "epoch": 0.5295750920321957, + "grad_norm": 0.25410348176956177, + "learning_rate": 0.00016469603377633577, + "loss": 0.1124, + "step": 33950 + }, + { + "epoch": 0.5297310788045174, + "grad_norm": 1.770642876625061, + "learning_rate": 0.000164685634658181, + "loss": 0.3295, + "step": 33960 + }, + { + "epoch": 0.5298870655768391, + "grad_norm": 0.0607205331325531, + "learning_rate": 0.0001646752355400262, + "loss": 0.2369, + "step": 33970 + }, + { + "epoch": 0.5300430523491608, + "grad_norm": 0.5557095408439636, + "learning_rate": 0.00016466483642187143, + "loss": 0.1681, + "step": 33980 + }, + { + "epoch": 0.5301990391214825, + "grad_norm": 0.5192957520484924, + "learning_rate": 0.00016465443730371665, + "loss": 0.2649, + "step": 33990 + }, + { + "epoch": 0.5303550258938042, + "grad_norm": 0.04804835096001625, + "learning_rate": 0.00016464403818556187, + "loss": 0.2639, + "step": 34000 + }, + { + "epoch": 0.5305110126661259, + "grad_norm": 0.02673129364848137, + "learning_rate": 0.0001646336390674071, + "loss": 0.169, + "step": 34010 + }, + { + "epoch": 0.5306669994384476, + "grad_norm": 1.0084244012832642, + "learning_rate": 0.0001646232399492523, + "loss": 0.353, + "step": 34020 + }, + { + "epoch": 0.5308229862107693, + "grad_norm": 2.2202091217041016, + "learning_rate": 0.00016461284083109753, + "loss": 0.3206, + "step": 34030 + }, + { + "epoch": 0.530978972983091, + "grad_norm": 0.5573744773864746, + "learning_rate": 0.00016460244171294275, + "loss": 0.231, + "step": 34040 + }, + { + "epoch": 0.5311349597554127, + "grad_norm": 0.39700084924697876, + "learning_rate": 0.00016459204259478797, + "loss": 0.1412, + "step": 34050 + }, + { + "epoch": 0.5312909465277345, + "grad_norm": 2.582963228225708, + "learning_rate": 0.00016458164347663319, + "loss": 0.3334, + "step": 34060 + }, + { + "epoch": 0.5314469333000561, + "grad_norm": 0.22781169414520264, + "learning_rate": 0.0001645712443584784, + "loss": 0.2767, + "step": 34070 + }, + { + "epoch": 0.5316029200723779, + "grad_norm": 1.2051042318344116, + "learning_rate": 0.00016456084524032362, + "loss": 0.2045, + "step": 34080 + }, + { + "epoch": 0.5317589068446996, + "grad_norm": 0.42760100960731506, + "learning_rate": 0.00016455044612216884, + "loss": 0.1151, + "step": 34090 + }, + { + "epoch": 0.5319148936170213, + "grad_norm": 0.14440476894378662, + "learning_rate": 0.00016454004700401406, + "loss": 0.2386, + "step": 34100 + }, + { + "epoch": 0.532070880389343, + "grad_norm": 2.2777981758117676, + "learning_rate": 0.00016452964788585928, + "loss": 0.2009, + "step": 34110 + }, + { + "epoch": 0.5322268671616647, + "grad_norm": 0.9206979274749756, + "learning_rate": 0.0001645192487677045, + "loss": 0.2745, + "step": 34120 + }, + { + "epoch": 0.5323828539339864, + "grad_norm": 1.6947574615478516, + "learning_rate": 0.00016450884964954972, + "loss": 0.2584, + "step": 34130 + }, + { + "epoch": 0.5325388407063081, + "grad_norm": 0.401444673538208, + "learning_rate": 0.00016449845053139494, + "loss": 0.2218, + "step": 34140 + }, + { + "epoch": 0.5326948274786298, + "grad_norm": 0.08261553198099136, + "learning_rate": 0.00016448805141324016, + "loss": 0.2775, + "step": 34150 + }, + { + "epoch": 0.5328508142509515, + "grad_norm": 0.1017974391579628, + "learning_rate": 0.00016447765229508538, + "loss": 0.2095, + "step": 34160 + }, + { + "epoch": 0.5330068010232732, + "grad_norm": 1.3759571313858032, + "learning_rate": 0.0001644672531769306, + "loss": 0.2643, + "step": 34170 + }, + { + "epoch": 0.5331627877955949, + "grad_norm": 1.2654389142990112, + "learning_rate": 0.00016445685405877582, + "loss": 0.2949, + "step": 34180 + }, + { + "epoch": 0.5333187745679167, + "grad_norm": 1.5481843948364258, + "learning_rate": 0.00016444645494062104, + "loss": 0.1264, + "step": 34190 + }, + { + "epoch": 0.5334747613402383, + "grad_norm": 1.8094528913497925, + "learning_rate": 0.00016443605582246626, + "loss": 0.2727, + "step": 34200 + }, + { + "epoch": 0.5336307481125601, + "grad_norm": 2.224538564682007, + "learning_rate": 0.00016442565670431148, + "loss": 0.3096, + "step": 34210 + }, + { + "epoch": 0.5337867348848817, + "grad_norm": 0.6375226974487305, + "learning_rate": 0.0001644152575861567, + "loss": 0.2251, + "step": 34220 + }, + { + "epoch": 0.5339427216572035, + "grad_norm": 3.727106809616089, + "learning_rate": 0.00016440485846800192, + "loss": 0.4374, + "step": 34230 + }, + { + "epoch": 0.5340987084295252, + "grad_norm": 0.13345426321029663, + "learning_rate": 0.00016439445934984714, + "loss": 0.2011, + "step": 34240 + }, + { + "epoch": 0.5342546952018469, + "grad_norm": 2.1658668518066406, + "learning_rate": 0.00016438406023169236, + "loss": 0.2457, + "step": 34250 + }, + { + "epoch": 0.5344106819741686, + "grad_norm": 19.238407135009766, + "learning_rate": 0.00016437366111353758, + "loss": 0.2756, + "step": 34260 + }, + { + "epoch": 0.5345666687464903, + "grad_norm": 1.0292778015136719, + "learning_rate": 0.0001643632619953828, + "loss": 0.1646, + "step": 34270 + }, + { + "epoch": 0.534722655518812, + "grad_norm": 0.9372987747192383, + "learning_rate": 0.00016435286287722802, + "loss": 0.2762, + "step": 34280 + }, + { + "epoch": 0.5348786422911337, + "grad_norm": 0.3918002247810364, + "learning_rate": 0.00016434246375907324, + "loss": 0.1406, + "step": 34290 + }, + { + "epoch": 0.5350346290634554, + "grad_norm": 1.3518732786178589, + "learning_rate": 0.00016433206464091846, + "loss": 0.2992, + "step": 34300 + }, + { + "epoch": 0.5351906158357771, + "grad_norm": 0.73117595911026, + "learning_rate": 0.00016432166552276368, + "loss": 0.1172, + "step": 34310 + }, + { + "epoch": 0.5353466026080989, + "grad_norm": 0.06655958294868469, + "learning_rate": 0.0001643112664046089, + "loss": 0.1332, + "step": 34320 + }, + { + "epoch": 0.5355025893804205, + "grad_norm": 3.2587168216705322, + "learning_rate": 0.00016430086728645412, + "loss": 0.2052, + "step": 34330 + }, + { + "epoch": 0.5356585761527423, + "grad_norm": 2.6668431758880615, + "learning_rate": 0.00016429046816829934, + "loss": 0.1295, + "step": 34340 + }, + { + "epoch": 0.5358145629250639, + "grad_norm": 1.316080927848816, + "learning_rate": 0.00016428006905014456, + "loss": 0.2072, + "step": 34350 + }, + { + "epoch": 0.5359705496973857, + "grad_norm": 5.144528388977051, + "learning_rate": 0.00016426966993198977, + "loss": 0.1934, + "step": 34360 + }, + { + "epoch": 0.5361265364697073, + "grad_norm": 1.3529599905014038, + "learning_rate": 0.000164259270813835, + "loss": 0.2077, + "step": 34370 + }, + { + "epoch": 0.5362825232420291, + "grad_norm": 0.9132925868034363, + "learning_rate": 0.00016424887169568021, + "loss": 0.2404, + "step": 34380 + }, + { + "epoch": 0.5364385100143508, + "grad_norm": 1.520033597946167, + "learning_rate": 0.00016423847257752543, + "loss": 0.1765, + "step": 34390 + }, + { + "epoch": 0.5365944967866725, + "grad_norm": 2.343975305557251, + "learning_rate": 0.00016422807345937065, + "loss": 0.1959, + "step": 34400 + }, + { + "epoch": 0.5367504835589942, + "grad_norm": 0.5855118632316589, + "learning_rate": 0.00016421767434121587, + "loss": 0.0848, + "step": 34410 + }, + { + "epoch": 0.5369064703313159, + "grad_norm": 1.9108648300170898, + "learning_rate": 0.0001642072752230611, + "loss": 0.2406, + "step": 34420 + }, + { + "epoch": 0.5370624571036376, + "grad_norm": 1.0323792695999146, + "learning_rate": 0.0001641968761049063, + "loss": 0.2298, + "step": 34430 + }, + { + "epoch": 0.5372184438759593, + "grad_norm": 0.17435620725154877, + "learning_rate": 0.00016418647698675153, + "loss": 0.1396, + "step": 34440 + }, + { + "epoch": 0.537374430648281, + "grad_norm": 0.3526577353477478, + "learning_rate": 0.00016417607786859675, + "loss": 0.172, + "step": 34450 + }, + { + "epoch": 0.5375304174206027, + "grad_norm": 2.73685884475708, + "learning_rate": 0.00016416567875044197, + "loss": 0.1167, + "step": 34460 + }, + { + "epoch": 0.5376864041929245, + "grad_norm": 2.765693426132202, + "learning_rate": 0.0001641552796322872, + "loss": 0.1356, + "step": 34470 + }, + { + "epoch": 0.5378423909652461, + "grad_norm": 0.47628021240234375, + "learning_rate": 0.0001641448805141324, + "loss": 0.3104, + "step": 34480 + }, + { + "epoch": 0.5379983777375679, + "grad_norm": 1.1307590007781982, + "learning_rate": 0.00016413448139597763, + "loss": 0.3247, + "step": 34490 + }, + { + "epoch": 0.5381543645098895, + "grad_norm": 0.07136381417512894, + "learning_rate": 0.00016412408227782285, + "loss": 0.319, + "step": 34500 + }, + { + "epoch": 0.5383103512822113, + "grad_norm": 1.7850221395492554, + "learning_rate": 0.00016411368315966807, + "loss": 0.2515, + "step": 34510 + }, + { + "epoch": 0.5384663380545329, + "grad_norm": 1.3870742321014404, + "learning_rate": 0.0001641032840415133, + "loss": 0.256, + "step": 34520 + }, + { + "epoch": 0.5386223248268547, + "grad_norm": 1.1329221725463867, + "learning_rate": 0.0001640928849233585, + "loss": 0.2255, + "step": 34530 + }, + { + "epoch": 0.5387783115991764, + "grad_norm": 0.23236137628555298, + "learning_rate": 0.00016408248580520373, + "loss": 0.4053, + "step": 34540 + }, + { + "epoch": 0.5389342983714981, + "grad_norm": 3.350924253463745, + "learning_rate": 0.00016407208668704895, + "loss": 0.2816, + "step": 34550 + }, + { + "epoch": 0.5390902851438198, + "grad_norm": 1.7459170818328857, + "learning_rate": 0.00016406168756889417, + "loss": 0.0834, + "step": 34560 + }, + { + "epoch": 0.5392462719161415, + "grad_norm": 0.89351487159729, + "learning_rate": 0.0001640512884507394, + "loss": 0.1146, + "step": 34570 + }, + { + "epoch": 0.5394022586884633, + "grad_norm": 0.49410831928253174, + "learning_rate": 0.0001640408893325846, + "loss": 0.2319, + "step": 34580 + }, + { + "epoch": 0.5395582454607849, + "grad_norm": 0.009971237741410732, + "learning_rate": 0.00016403049021442983, + "loss": 0.1674, + "step": 34590 + }, + { + "epoch": 0.5397142322331067, + "grad_norm": 1.420511245727539, + "learning_rate": 0.00016402009109627505, + "loss": 0.1469, + "step": 34600 + }, + { + "epoch": 0.5398702190054283, + "grad_norm": 2.5686769485473633, + "learning_rate": 0.00016400969197812027, + "loss": 0.2607, + "step": 34610 + }, + { + "epoch": 0.5400262057777501, + "grad_norm": 2.5808913707733154, + "learning_rate": 0.00016399929285996549, + "loss": 0.4412, + "step": 34620 + }, + { + "epoch": 0.5401821925500717, + "grad_norm": 0.2778591811656952, + "learning_rate": 0.0001639888937418107, + "loss": 0.1563, + "step": 34630 + }, + { + "epoch": 0.5403381793223935, + "grad_norm": 1.6448099613189697, + "learning_rate": 0.00016397849462365592, + "loss": 0.2484, + "step": 34640 + }, + { + "epoch": 0.5404941660947151, + "grad_norm": 0.003213417250663042, + "learning_rate": 0.00016396809550550114, + "loss": 0.1582, + "step": 34650 + }, + { + "epoch": 0.5406501528670369, + "grad_norm": 2.3433990478515625, + "learning_rate": 0.00016395769638734636, + "loss": 0.2774, + "step": 34660 + }, + { + "epoch": 0.5408061396393585, + "grad_norm": 0.09544426947832108, + "learning_rate": 0.00016394729726919158, + "loss": 0.1242, + "step": 34670 + }, + { + "epoch": 0.5409621264116803, + "grad_norm": 1.8934235572814941, + "learning_rate": 0.0001639368981510368, + "loss": 0.1636, + "step": 34680 + }, + { + "epoch": 0.541118113184002, + "grad_norm": 2.8528554439544678, + "learning_rate": 0.00016392649903288202, + "loss": 0.2524, + "step": 34690 + }, + { + "epoch": 0.5412740999563237, + "grad_norm": 2.837836980819702, + "learning_rate": 0.00016391609991472722, + "loss": 0.1808, + "step": 34700 + }, + { + "epoch": 0.5414300867286455, + "grad_norm": 4.349740028381348, + "learning_rate": 0.00016390570079657246, + "loss": 0.12, + "step": 34710 + }, + { + "epoch": 0.5415860735009671, + "grad_norm": 0.6011945605278015, + "learning_rate": 0.00016389530167841765, + "loss": 0.2865, + "step": 34720 + }, + { + "epoch": 0.5417420602732889, + "grad_norm": 1.2079488039016724, + "learning_rate": 0.0001638849025602629, + "loss": 0.125, + "step": 34730 + }, + { + "epoch": 0.5418980470456105, + "grad_norm": 1.9549164772033691, + "learning_rate": 0.0001638745034421081, + "loss": 0.3285, + "step": 34740 + }, + { + "epoch": 0.5420540338179323, + "grad_norm": 0.6969407200813293, + "learning_rate": 0.00016386410432395334, + "loss": 0.3267, + "step": 34750 + }, + { + "epoch": 0.5422100205902539, + "grad_norm": 1.0561965703964233, + "learning_rate": 0.00016385370520579853, + "loss": 0.1463, + "step": 34760 + }, + { + "epoch": 0.5423660073625757, + "grad_norm": 1.8208931684494019, + "learning_rate": 0.00016384330608764378, + "loss": 0.1713, + "step": 34770 + }, + { + "epoch": 0.5425219941348973, + "grad_norm": 0.6349910497665405, + "learning_rate": 0.000163832906969489, + "loss": 0.2244, + "step": 34780 + }, + { + "epoch": 0.5426779809072191, + "grad_norm": 1.1976515054702759, + "learning_rate": 0.00016382250785133422, + "loss": 0.3407, + "step": 34790 + }, + { + "epoch": 0.5428339676795407, + "grad_norm": 0.5970319509506226, + "learning_rate": 0.00016381210873317944, + "loss": 0.1419, + "step": 34800 + }, + { + "epoch": 0.5429899544518625, + "grad_norm": 0.14623159170150757, + "learning_rate": 0.00016380170961502466, + "loss": 0.14, + "step": 34810 + }, + { + "epoch": 0.5431459412241841, + "grad_norm": 1.4332351684570312, + "learning_rate": 0.00016379131049686988, + "loss": 0.1093, + "step": 34820 + }, + { + "epoch": 0.5433019279965059, + "grad_norm": 1.1443063020706177, + "learning_rate": 0.0001637809113787151, + "loss": 0.3676, + "step": 34830 + }, + { + "epoch": 0.5434579147688277, + "grad_norm": 1.3897461891174316, + "learning_rate": 0.00016377051226056032, + "loss": 0.3713, + "step": 34840 + }, + { + "epoch": 0.5436139015411493, + "grad_norm": 1.7773199081420898, + "learning_rate": 0.00016376011314240554, + "loss": 0.2249, + "step": 34850 + }, + { + "epoch": 0.5437698883134711, + "grad_norm": 0.8524389266967773, + "learning_rate": 0.00016374971402425076, + "loss": 0.2158, + "step": 34860 + }, + { + "epoch": 0.5439258750857927, + "grad_norm": 0.5572389960289001, + "learning_rate": 0.00016373931490609598, + "loss": 0.215, + "step": 34870 + }, + { + "epoch": 0.5440818618581145, + "grad_norm": 3.0238709449768066, + "learning_rate": 0.0001637289157879412, + "loss": 0.3765, + "step": 34880 + }, + { + "epoch": 0.5442378486304361, + "grad_norm": 1.1655421257019043, + "learning_rate": 0.00016371851666978642, + "loss": 0.2491, + "step": 34890 + }, + { + "epoch": 0.5443938354027579, + "grad_norm": 0.49492090940475464, + "learning_rate": 0.00016370811755163164, + "loss": 0.1287, + "step": 34900 + }, + { + "epoch": 0.5445498221750795, + "grad_norm": 0.2732921540737152, + "learning_rate": 0.00016369771843347685, + "loss": 0.2154, + "step": 34910 + }, + { + "epoch": 0.5447058089474013, + "grad_norm": 2.5807769298553467, + "learning_rate": 0.00016368731931532207, + "loss": 0.2032, + "step": 34920 + }, + { + "epoch": 0.5448617957197229, + "grad_norm": 0.5044315457344055, + "learning_rate": 0.0001636769201971673, + "loss": 0.2595, + "step": 34930 + }, + { + "epoch": 0.5450177824920447, + "grad_norm": 2.1248972415924072, + "learning_rate": 0.00016366652107901251, + "loss": 0.5651, + "step": 34940 + }, + { + "epoch": 0.5451737692643663, + "grad_norm": 0.8391468524932861, + "learning_rate": 0.00016365612196085773, + "loss": 0.3028, + "step": 34950 + }, + { + "epoch": 0.5453297560366881, + "grad_norm": 0.36081477999687195, + "learning_rate": 0.00016364572284270295, + "loss": 0.1073, + "step": 34960 + }, + { + "epoch": 0.5454857428090097, + "grad_norm": 3.175804376602173, + "learning_rate": 0.00016363532372454817, + "loss": 0.1889, + "step": 34970 + }, + { + "epoch": 0.5456417295813315, + "grad_norm": 0.20647937059402466, + "learning_rate": 0.0001636249246063934, + "loss": 0.2416, + "step": 34980 + }, + { + "epoch": 0.5457977163536533, + "grad_norm": 2.3525943756103516, + "learning_rate": 0.0001636145254882386, + "loss": 0.3179, + "step": 34990 + }, + { + "epoch": 0.5459537031259749, + "grad_norm": 0.2641216814517975, + "learning_rate": 0.00016360412637008383, + "loss": 0.2157, + "step": 35000 + }, + { + "epoch": 0.5461096898982967, + "grad_norm": 0.09906073659658432, + "learning_rate": 0.00016359372725192905, + "loss": 0.2576, + "step": 35010 + }, + { + "epoch": 0.5462656766706183, + "grad_norm": 0.13971920311450958, + "learning_rate": 0.00016358332813377427, + "loss": 0.1298, + "step": 35020 + }, + { + "epoch": 0.5464216634429401, + "grad_norm": 1.4535638093948364, + "learning_rate": 0.0001635729290156195, + "loss": 0.2357, + "step": 35030 + }, + { + "epoch": 0.5465776502152617, + "grad_norm": 0.9107828736305237, + "learning_rate": 0.0001635625298974647, + "loss": 0.3049, + "step": 35040 + }, + { + "epoch": 0.5467336369875835, + "grad_norm": 0.0377386212348938, + "learning_rate": 0.00016355213077930993, + "loss": 0.2066, + "step": 35050 + }, + { + "epoch": 0.5468896237599051, + "grad_norm": 2.7800869941711426, + "learning_rate": 0.00016354173166115515, + "loss": 0.2974, + "step": 35060 + }, + { + "epoch": 0.5470456105322269, + "grad_norm": 1.6247998476028442, + "learning_rate": 0.00016353133254300037, + "loss": 0.3822, + "step": 35070 + }, + { + "epoch": 0.5472015973045485, + "grad_norm": 1.6479015350341797, + "learning_rate": 0.0001635209334248456, + "loss": 0.3401, + "step": 35080 + }, + { + "epoch": 0.5473575840768703, + "grad_norm": 0.1879737675189972, + "learning_rate": 0.0001635105343066908, + "loss": 0.1549, + "step": 35090 + }, + { + "epoch": 0.5475135708491919, + "grad_norm": 0.39355361461639404, + "learning_rate": 0.00016350013518853603, + "loss": 0.2032, + "step": 35100 + }, + { + "epoch": 0.5476695576215137, + "grad_norm": 2.9976983070373535, + "learning_rate": 0.00016348973607038125, + "loss": 0.3867, + "step": 35110 + }, + { + "epoch": 0.5478255443938354, + "grad_norm": 0.8242707252502441, + "learning_rate": 0.00016347933695222647, + "loss": 0.3601, + "step": 35120 + }, + { + "epoch": 0.5479815311661571, + "grad_norm": 0.874955415725708, + "learning_rate": 0.0001634689378340717, + "loss": 0.2709, + "step": 35130 + }, + { + "epoch": 0.5481375179384789, + "grad_norm": 0.05980971083045006, + "learning_rate": 0.0001634585387159169, + "loss": 0.2646, + "step": 35140 + }, + { + "epoch": 0.5482935047108005, + "grad_norm": 1.8437328338623047, + "learning_rate": 0.0001634481395977621, + "loss": 0.1965, + "step": 35150 + }, + { + "epoch": 0.5484494914831223, + "grad_norm": 0.4142405688762665, + "learning_rate": 0.00016343774047960735, + "loss": 0.1756, + "step": 35160 + }, + { + "epoch": 0.5486054782554439, + "grad_norm": 4.836324214935303, + "learning_rate": 0.00016342734136145254, + "loss": 0.2534, + "step": 35170 + }, + { + "epoch": 0.5487614650277657, + "grad_norm": 1.4933065176010132, + "learning_rate": 0.00016341694224329779, + "loss": 0.2355, + "step": 35180 + }, + { + "epoch": 0.5489174518000873, + "grad_norm": 1.9468894004821777, + "learning_rate": 0.00016340654312514298, + "loss": 0.3113, + "step": 35190 + }, + { + "epoch": 0.5490734385724091, + "grad_norm": 1.012710452079773, + "learning_rate": 0.00016339614400698822, + "loss": 0.2635, + "step": 35200 + }, + { + "epoch": 0.5492294253447307, + "grad_norm": 1.5551140308380127, + "learning_rate": 0.00016338574488883342, + "loss": 0.2951, + "step": 35210 + }, + { + "epoch": 0.5493854121170525, + "grad_norm": 1.8613696098327637, + "learning_rate": 0.00016337534577067866, + "loss": 0.2832, + "step": 35220 + }, + { + "epoch": 0.5495413988893741, + "grad_norm": 1.6748839616775513, + "learning_rate": 0.00016336494665252386, + "loss": 0.2266, + "step": 35230 + }, + { + "epoch": 0.5496973856616959, + "grad_norm": 0.4094032347202301, + "learning_rate": 0.0001633545475343691, + "loss": 0.3497, + "step": 35240 + }, + { + "epoch": 0.5498533724340176, + "grad_norm": 1.6414631605148315, + "learning_rate": 0.0001633441484162143, + "loss": 0.1763, + "step": 35250 + }, + { + "epoch": 0.5500093592063393, + "grad_norm": 1.6380645036697388, + "learning_rate": 0.00016333374929805954, + "loss": 0.1355, + "step": 35260 + }, + { + "epoch": 0.550165345978661, + "grad_norm": 0.7325630187988281, + "learning_rate": 0.00016332335017990473, + "loss": 0.3105, + "step": 35270 + }, + { + "epoch": 0.5503213327509827, + "grad_norm": 1.6015644073486328, + "learning_rate": 0.00016331295106174998, + "loss": 0.2311, + "step": 35280 + }, + { + "epoch": 0.5504773195233045, + "grad_norm": 0.9224210381507874, + "learning_rate": 0.00016330255194359517, + "loss": 0.2746, + "step": 35290 + }, + { + "epoch": 0.5506333062956261, + "grad_norm": 1.30025315284729, + "learning_rate": 0.00016329215282544042, + "loss": 0.2066, + "step": 35300 + }, + { + "epoch": 0.5507892930679479, + "grad_norm": 0.481125146150589, + "learning_rate": 0.0001632817537072856, + "loss": 0.1778, + "step": 35310 + }, + { + "epoch": 0.5509452798402695, + "grad_norm": 0.2709486782550812, + "learning_rate": 0.00016327135458913086, + "loss": 0.1131, + "step": 35320 + }, + { + "epoch": 0.5511012666125913, + "grad_norm": 0.6248563528060913, + "learning_rate": 0.00016326095547097605, + "loss": 0.4356, + "step": 35330 + }, + { + "epoch": 0.5512572533849129, + "grad_norm": 3.0947647094726562, + "learning_rate": 0.0001632505563528213, + "loss": 0.3072, + "step": 35340 + }, + { + "epoch": 0.5514132401572347, + "grad_norm": 1.009535312652588, + "learning_rate": 0.0001632401572346665, + "loss": 0.1808, + "step": 35350 + }, + { + "epoch": 0.5515692269295563, + "grad_norm": 0.042604975402355194, + "learning_rate": 0.00016322975811651174, + "loss": 0.1198, + "step": 35360 + }, + { + "epoch": 0.5517252137018781, + "grad_norm": 1.8652396202087402, + "learning_rate": 0.00016321935899835693, + "loss": 0.2176, + "step": 35370 + }, + { + "epoch": 0.5518812004741998, + "grad_norm": 4.6097187995910645, + "learning_rate": 0.00016320895988020218, + "loss": 0.2132, + "step": 35380 + }, + { + "epoch": 0.5520371872465215, + "grad_norm": 1.3615522384643555, + "learning_rate": 0.00016319856076204737, + "loss": 0.2751, + "step": 35390 + }, + { + "epoch": 0.5521931740188432, + "grad_norm": 0.8844773769378662, + "learning_rate": 0.00016318816164389262, + "loss": 0.22, + "step": 35400 + }, + { + "epoch": 0.5523491607911649, + "grad_norm": 0.5413331985473633, + "learning_rate": 0.0001631777625257378, + "loss": 0.2159, + "step": 35410 + }, + { + "epoch": 0.5525051475634866, + "grad_norm": 1.5456678867340088, + "learning_rate": 0.00016316736340758306, + "loss": 0.2244, + "step": 35420 + }, + { + "epoch": 0.5526611343358083, + "grad_norm": 2.1405861377716064, + "learning_rate": 0.00016315696428942825, + "loss": 0.446, + "step": 35430 + }, + { + "epoch": 0.5528171211081301, + "grad_norm": 1.4269858598709106, + "learning_rate": 0.0001631465651712735, + "loss": 0.3599, + "step": 35440 + }, + { + "epoch": 0.5529731078804517, + "grad_norm": 1.8682516813278198, + "learning_rate": 0.0001631361660531187, + "loss": 0.115, + "step": 35450 + }, + { + "epoch": 0.5531290946527735, + "grad_norm": 1.0175774097442627, + "learning_rate": 0.00016312576693496394, + "loss": 0.4003, + "step": 35460 + }, + { + "epoch": 0.5532850814250951, + "grad_norm": 2.4369170665740967, + "learning_rate": 0.00016311536781680913, + "loss": 0.2489, + "step": 35470 + }, + { + "epoch": 0.5534410681974169, + "grad_norm": 1.886022686958313, + "learning_rate": 0.00016310496869865437, + "loss": 0.4201, + "step": 35480 + }, + { + "epoch": 0.5535970549697385, + "grad_norm": 1.5811959505081177, + "learning_rate": 0.00016309456958049957, + "loss": 0.4286, + "step": 35490 + }, + { + "epoch": 0.5537530417420603, + "grad_norm": 1.971110224723816, + "learning_rate": 0.00016308417046234481, + "loss": 0.3599, + "step": 35500 + }, + { + "epoch": 0.553909028514382, + "grad_norm": 0.0335380844771862, + "learning_rate": 0.00016307377134419, + "loss": 0.3357, + "step": 35510 + }, + { + "epoch": 0.5540650152867037, + "grad_norm": 1.8519577980041504, + "learning_rate": 0.00016306337222603525, + "loss": 0.3126, + "step": 35520 + }, + { + "epoch": 0.5542210020590254, + "grad_norm": 2.1563379764556885, + "learning_rate": 0.00016305297310788045, + "loss": 0.1123, + "step": 35530 + }, + { + "epoch": 0.5543769888313471, + "grad_norm": 1.64332914352417, + "learning_rate": 0.0001630425739897257, + "loss": 0.2853, + "step": 35540 + }, + { + "epoch": 0.5545329756036688, + "grad_norm": 0.061150554567575455, + "learning_rate": 0.00016303217487157088, + "loss": 0.1943, + "step": 35550 + }, + { + "epoch": 0.5546889623759905, + "grad_norm": 1.2701060771942139, + "learning_rate": 0.00016302177575341613, + "loss": 0.2304, + "step": 35560 + }, + { + "epoch": 0.5548449491483122, + "grad_norm": 2.424860715866089, + "learning_rate": 0.00016301137663526132, + "loss": 0.2128, + "step": 35570 + }, + { + "epoch": 0.5550009359206339, + "grad_norm": 0.6803575158119202, + "learning_rate": 0.00016300097751710657, + "loss": 0.1145, + "step": 35580 + }, + { + "epoch": 0.5551569226929557, + "grad_norm": 1.2855092287063599, + "learning_rate": 0.00016299057839895176, + "loss": 0.2448, + "step": 35590 + }, + { + "epoch": 0.5553129094652773, + "grad_norm": 2.6340911388397217, + "learning_rate": 0.00016298017928079698, + "loss": 0.1856, + "step": 35600 + }, + { + "epoch": 0.5554688962375991, + "grad_norm": 1.776382327079773, + "learning_rate": 0.0001629697801626422, + "loss": 0.2422, + "step": 35610 + }, + { + "epoch": 0.5556248830099207, + "grad_norm": 3.0746347904205322, + "learning_rate": 0.00016295938104448742, + "loss": 0.1578, + "step": 35620 + }, + { + "epoch": 0.5557808697822425, + "grad_norm": 1.4887659549713135, + "learning_rate": 0.00016294898192633267, + "loss": 0.2385, + "step": 35630 + }, + { + "epoch": 0.5559368565545642, + "grad_norm": 0.5371220707893372, + "learning_rate": 0.00016293858280817786, + "loss": 0.2361, + "step": 35640 + }, + { + "epoch": 0.5560928433268859, + "grad_norm": 0.18604904413223267, + "learning_rate": 0.0001629281836900231, + "loss": 0.0828, + "step": 35650 + }, + { + "epoch": 0.5562488300992076, + "grad_norm": 2.267854690551758, + "learning_rate": 0.0001629177845718683, + "loss": 0.413, + "step": 35660 + }, + { + "epoch": 0.5564048168715293, + "grad_norm": 1.9755452871322632, + "learning_rate": 0.00016290738545371355, + "loss": 0.146, + "step": 35670 + }, + { + "epoch": 0.556560803643851, + "grad_norm": 2.430293321609497, + "learning_rate": 0.00016289698633555874, + "loss": 0.2677, + "step": 35680 + }, + { + "epoch": 0.5567167904161727, + "grad_norm": 0.6319543719291687, + "learning_rate": 0.000162886587217404, + "loss": 0.1481, + "step": 35690 + }, + { + "epoch": 0.5568727771884944, + "grad_norm": 0.17439277470111847, + "learning_rate": 0.00016287618809924918, + "loss": 0.1641, + "step": 35700 + }, + { + "epoch": 0.5570287639608161, + "grad_norm": 0.350175142288208, + "learning_rate": 0.00016286578898109443, + "loss": 0.393, + "step": 35710 + }, + { + "epoch": 0.5571847507331378, + "grad_norm": 2.641941547393799, + "learning_rate": 0.00016285538986293962, + "loss": 0.1907, + "step": 35720 + }, + { + "epoch": 0.5573407375054595, + "grad_norm": 0.1910303384065628, + "learning_rate": 0.00016284499074478487, + "loss": 0.2248, + "step": 35730 + }, + { + "epoch": 0.5574967242777813, + "grad_norm": 1.0180896520614624, + "learning_rate": 0.00016283459162663006, + "loss": 0.1924, + "step": 35740 + }, + { + "epoch": 0.557652711050103, + "grad_norm": 1.2047260999679565, + "learning_rate": 0.0001628241925084753, + "loss": 0.2031, + "step": 35750 + }, + { + "epoch": 0.5578086978224247, + "grad_norm": 1.8182405233383179, + "learning_rate": 0.0001628137933903205, + "loss": 0.1573, + "step": 35760 + }, + { + "epoch": 0.5579646845947464, + "grad_norm": 0.5485963821411133, + "learning_rate": 0.00016280339427216574, + "loss": 0.152, + "step": 35770 + }, + { + "epoch": 0.5581206713670681, + "grad_norm": 0.1674145758152008, + "learning_rate": 0.00016279299515401094, + "loss": 0.2388, + "step": 35780 + }, + { + "epoch": 0.5582766581393898, + "grad_norm": 0.36187657713890076, + "learning_rate": 0.00016278259603585618, + "loss": 0.1618, + "step": 35790 + }, + { + "epoch": 0.5584326449117115, + "grad_norm": 0.013768521137535572, + "learning_rate": 0.00016277219691770138, + "loss": 0.3666, + "step": 35800 + }, + { + "epoch": 0.5585886316840332, + "grad_norm": 2.8602802753448486, + "learning_rate": 0.00016276179779954662, + "loss": 0.2065, + "step": 35810 + }, + { + "epoch": 0.5587446184563549, + "grad_norm": 1.7925455570220947, + "learning_rate": 0.00016275139868139182, + "loss": 0.1729, + "step": 35820 + }, + { + "epoch": 0.5589006052286766, + "grad_norm": 2.910456418991089, + "learning_rate": 0.00016274099956323706, + "loss": 0.3734, + "step": 35830 + }, + { + "epoch": 0.5590565920009983, + "grad_norm": 0.8378308415412903, + "learning_rate": 0.00016273060044508225, + "loss": 0.2255, + "step": 35840 + }, + { + "epoch": 0.55921257877332, + "grad_norm": 0.409534752368927, + "learning_rate": 0.0001627202013269275, + "loss": 0.2046, + "step": 35850 + }, + { + "epoch": 0.5593685655456417, + "grad_norm": 0.49498099088668823, + "learning_rate": 0.0001627098022087727, + "loss": 0.1058, + "step": 35860 + }, + { + "epoch": 0.5595245523179634, + "grad_norm": 1.1761783361434937, + "learning_rate": 0.00016269940309061794, + "loss": 0.1566, + "step": 35870 + }, + { + "epoch": 0.5596805390902851, + "grad_norm": 3.8204751014709473, + "learning_rate": 0.00016268900397246313, + "loss": 0.2646, + "step": 35880 + }, + { + "epoch": 0.5598365258626069, + "grad_norm": 0.9882522225379944, + "learning_rate": 0.00016267860485430838, + "loss": 0.1756, + "step": 35890 + }, + { + "epoch": 0.5599925126349286, + "grad_norm": 1.1832259893417358, + "learning_rate": 0.00016266820573615357, + "loss": 0.1385, + "step": 35900 + }, + { + "epoch": 0.5601484994072503, + "grad_norm": 0.7638296484947205, + "learning_rate": 0.00016265780661799882, + "loss": 0.5679, + "step": 35910 + }, + { + "epoch": 0.560304486179572, + "grad_norm": 0.8551504611968994, + "learning_rate": 0.000162647407499844, + "loss": 0.2981, + "step": 35920 + }, + { + "epoch": 0.5604604729518937, + "grad_norm": 5.499948501586914, + "learning_rate": 0.00016263700838168926, + "loss": 0.1429, + "step": 35930 + }, + { + "epoch": 0.5606164597242154, + "grad_norm": 3.0494496822357178, + "learning_rate": 0.00016262660926353445, + "loss": 0.1322, + "step": 35940 + }, + { + "epoch": 0.5607724464965371, + "grad_norm": 0.2482384592294693, + "learning_rate": 0.0001626162101453797, + "loss": 0.2473, + "step": 35950 + }, + { + "epoch": 0.5609284332688588, + "grad_norm": 0.009750776924192905, + "learning_rate": 0.0001626058110272249, + "loss": 0.0429, + "step": 35960 + }, + { + "epoch": 0.5610844200411805, + "grad_norm": 0.5742604732513428, + "learning_rate": 0.00016259541190907014, + "loss": 0.4565, + "step": 35970 + }, + { + "epoch": 0.5612404068135022, + "grad_norm": 0.7810243368148804, + "learning_rate": 0.00016258501279091533, + "loss": 0.2107, + "step": 35980 + }, + { + "epoch": 0.5613963935858239, + "grad_norm": 2.785747528076172, + "learning_rate": 0.00016257461367276058, + "loss": 0.5087, + "step": 35990 + }, + { + "epoch": 0.5615523803581456, + "grad_norm": 0.8806902766227722, + "learning_rate": 0.00016256421455460577, + "loss": 0.2937, + "step": 36000 + }, + { + "epoch": 0.5617083671304673, + "grad_norm": 1.857373833656311, + "learning_rate": 0.00016255381543645102, + "loss": 0.5042, + "step": 36010 + }, + { + "epoch": 0.561864353902789, + "grad_norm": 0.5333901047706604, + "learning_rate": 0.0001625434163182962, + "loss": 0.1678, + "step": 36020 + }, + { + "epoch": 0.5620203406751108, + "grad_norm": 11.023160934448242, + "learning_rate": 0.00016253301720014145, + "loss": 0.2721, + "step": 36030 + }, + { + "epoch": 0.5621763274474325, + "grad_norm": 2.6214029788970947, + "learning_rate": 0.00016252261808198665, + "loss": 0.1675, + "step": 36040 + }, + { + "epoch": 0.5623323142197542, + "grad_norm": 0.25980064272880554, + "learning_rate": 0.0001625122189638319, + "loss": 0.1832, + "step": 36050 + }, + { + "epoch": 0.5624883009920759, + "grad_norm": 1.3559473752975464, + "learning_rate": 0.0001625018198456771, + "loss": 0.121, + "step": 36060 + }, + { + "epoch": 0.5626442877643976, + "grad_norm": 0.3558153212070465, + "learning_rate": 0.0001624914207275223, + "loss": 0.2797, + "step": 36070 + }, + { + "epoch": 0.5628002745367193, + "grad_norm": 0.6314427256584167, + "learning_rate": 0.00016248102160936753, + "loss": 0.3455, + "step": 36080 + }, + { + "epoch": 0.562956261309041, + "grad_norm": 0.16120101511478424, + "learning_rate": 0.00016247062249121275, + "loss": 0.3669, + "step": 36090 + }, + { + "epoch": 0.5631122480813627, + "grad_norm": 1.041572093963623, + "learning_rate": 0.00016246022337305797, + "loss": 0.2744, + "step": 36100 + }, + { + "epoch": 0.5632682348536844, + "grad_norm": 0.2645890414714813, + "learning_rate": 0.00016244982425490318, + "loss": 0.1487, + "step": 36110 + }, + { + "epoch": 0.5634242216260061, + "grad_norm": 5.698698043823242, + "learning_rate": 0.0001624394251367484, + "loss": 0.173, + "step": 36120 + }, + { + "epoch": 0.5635802083983278, + "grad_norm": 0.3299804627895355, + "learning_rate": 0.00016242902601859362, + "loss": 0.1965, + "step": 36130 + }, + { + "epoch": 0.5637361951706495, + "grad_norm": 1.2476481199264526, + "learning_rate": 0.00016241862690043884, + "loss": 0.4097, + "step": 36140 + }, + { + "epoch": 0.5638921819429712, + "grad_norm": 0.30926263332366943, + "learning_rate": 0.00016240822778228406, + "loss": 0.1416, + "step": 36150 + }, + { + "epoch": 0.564048168715293, + "grad_norm": 0.05872740224003792, + "learning_rate": 0.00016239782866412928, + "loss": 0.1449, + "step": 36160 + }, + { + "epoch": 0.5642041554876146, + "grad_norm": 0.2292211800813675, + "learning_rate": 0.0001623874295459745, + "loss": 0.2557, + "step": 36170 + }, + { + "epoch": 0.5643601422599364, + "grad_norm": 1.7822531461715698, + "learning_rate": 0.00016237703042781972, + "loss": 0.2953, + "step": 36180 + }, + { + "epoch": 0.5645161290322581, + "grad_norm": 2.4908461570739746, + "learning_rate": 0.00016236663130966494, + "loss": 0.3888, + "step": 36190 + }, + { + "epoch": 0.5646721158045798, + "grad_norm": 0.08363594114780426, + "learning_rate": 0.00016235623219151016, + "loss": 0.2293, + "step": 36200 + }, + { + "epoch": 0.5648281025769015, + "grad_norm": 1.4830002784729004, + "learning_rate": 0.00016234583307335538, + "loss": 0.2348, + "step": 36210 + }, + { + "epoch": 0.5649840893492232, + "grad_norm": 1.8443071842193604, + "learning_rate": 0.0001623354339552006, + "loss": 0.3392, + "step": 36220 + }, + { + "epoch": 0.5651400761215449, + "grad_norm": 2.8050875663757324, + "learning_rate": 0.00016232503483704582, + "loss": 0.2426, + "step": 36230 + }, + { + "epoch": 0.5652960628938666, + "grad_norm": 3.6627259254455566, + "learning_rate": 0.00016231463571889104, + "loss": 0.2118, + "step": 36240 + }, + { + "epoch": 0.5654520496661883, + "grad_norm": 0.3443094789981842, + "learning_rate": 0.00016230423660073626, + "loss": 0.2418, + "step": 36250 + }, + { + "epoch": 0.56560803643851, + "grad_norm": 1.7411112785339355, + "learning_rate": 0.00016229383748258148, + "loss": 0.1628, + "step": 36260 + }, + { + "epoch": 0.5657640232108317, + "grad_norm": 1.1321816444396973, + "learning_rate": 0.0001622834383644267, + "loss": 0.2732, + "step": 36270 + }, + { + "epoch": 0.5659200099831534, + "grad_norm": 2.9686615467071533, + "learning_rate": 0.00016227303924627192, + "loss": 0.3036, + "step": 36280 + }, + { + "epoch": 0.5660759967554752, + "grad_norm": 5.03530740737915, + "learning_rate": 0.00016226264012811714, + "loss": 0.2143, + "step": 36290 + }, + { + "epoch": 0.5662319835277968, + "grad_norm": 2.3943281173706055, + "learning_rate": 0.00016225224100996236, + "loss": 0.2629, + "step": 36300 + }, + { + "epoch": 0.5663879703001186, + "grad_norm": 1.812828779220581, + "learning_rate": 0.00016224184189180758, + "loss": 0.3892, + "step": 36310 + }, + { + "epoch": 0.5665439570724402, + "grad_norm": 1.4708483219146729, + "learning_rate": 0.0001622314427736528, + "loss": 0.2118, + "step": 36320 + }, + { + "epoch": 0.566699943844762, + "grad_norm": 8.99713134765625, + "learning_rate": 0.00016222104365549802, + "loss": 0.3185, + "step": 36330 + }, + { + "epoch": 0.5668559306170837, + "grad_norm": 1.7472341060638428, + "learning_rate": 0.00016221064453734324, + "loss": 0.1659, + "step": 36340 + }, + { + "epoch": 0.5670119173894054, + "grad_norm": 0.8059778213500977, + "learning_rate": 0.00016220024541918846, + "loss": 0.0621, + "step": 36350 + }, + { + "epoch": 0.5671679041617271, + "grad_norm": 4.0501861572265625, + "learning_rate": 0.00016218984630103368, + "loss": 0.1399, + "step": 36360 + }, + { + "epoch": 0.5673238909340488, + "grad_norm": 0.014400321058928967, + "learning_rate": 0.0001621794471828789, + "loss": 0.2372, + "step": 36370 + }, + { + "epoch": 0.5674798777063705, + "grad_norm": 0.4061933755874634, + "learning_rate": 0.00016216904806472412, + "loss": 0.3854, + "step": 36380 + }, + { + "epoch": 0.5676358644786922, + "grad_norm": 0.4940955340862274, + "learning_rate": 0.00016215864894656933, + "loss": 0.1689, + "step": 36390 + }, + { + "epoch": 0.567791851251014, + "grad_norm": 0.3220517635345459, + "learning_rate": 0.00016214824982841455, + "loss": 0.1912, + "step": 36400 + }, + { + "epoch": 0.5679478380233356, + "grad_norm": 2.0848920345306396, + "learning_rate": 0.00016213785071025977, + "loss": 0.1639, + "step": 36410 + }, + { + "epoch": 0.5681038247956574, + "grad_norm": 0.04129406809806824, + "learning_rate": 0.000162127451592105, + "loss": 0.1378, + "step": 36420 + }, + { + "epoch": 0.568259811567979, + "grad_norm": 1.9343401193618774, + "learning_rate": 0.0001621170524739502, + "loss": 0.3759, + "step": 36430 + }, + { + "epoch": 0.5684157983403008, + "grad_norm": 1.0625615119934082, + "learning_rate": 0.00016210665335579543, + "loss": 0.2074, + "step": 36440 + }, + { + "epoch": 0.5685717851126224, + "grad_norm": 0.14784491062164307, + "learning_rate": 0.00016209625423764065, + "loss": 0.0981, + "step": 36450 + }, + { + "epoch": 0.5687277718849442, + "grad_norm": 1.8117821216583252, + "learning_rate": 0.00016208585511948587, + "loss": 0.2755, + "step": 36460 + }, + { + "epoch": 0.5688837586572658, + "grad_norm": 0.061010006815195084, + "learning_rate": 0.0001620754560013311, + "loss": 0.1815, + "step": 36470 + }, + { + "epoch": 0.5690397454295876, + "grad_norm": 0.44453224539756775, + "learning_rate": 0.0001620650568831763, + "loss": 0.0891, + "step": 36480 + }, + { + "epoch": 0.5691957322019093, + "grad_norm": 0.6130645871162415, + "learning_rate": 0.00016205465776502153, + "loss": 0.2778, + "step": 36490 + }, + { + "epoch": 0.569351718974231, + "grad_norm": 5.408796787261963, + "learning_rate": 0.00016204425864686678, + "loss": 0.2301, + "step": 36500 + }, + { + "epoch": 0.5695077057465527, + "grad_norm": 1.2095412015914917, + "learning_rate": 0.00016203385952871197, + "loss": 0.1378, + "step": 36510 + }, + { + "epoch": 0.5696636925188744, + "grad_norm": 0.46873530745506287, + "learning_rate": 0.0001620234604105572, + "loss": 0.1431, + "step": 36520 + }, + { + "epoch": 0.5698196792911961, + "grad_norm": 1.028438925743103, + "learning_rate": 0.0001620130612924024, + "loss": 0.1114, + "step": 36530 + }, + { + "epoch": 0.5699756660635178, + "grad_norm": 0.5952587127685547, + "learning_rate": 0.00016200266217424763, + "loss": 0.2156, + "step": 36540 + }, + { + "epoch": 0.5701316528358396, + "grad_norm": 0.5134285688400269, + "learning_rate": 0.00016199226305609285, + "loss": 0.2712, + "step": 36550 + }, + { + "epoch": 0.5702876396081612, + "grad_norm": 0.03440163657069206, + "learning_rate": 0.00016198186393793807, + "loss": 0.312, + "step": 36560 + }, + { + "epoch": 0.570443626380483, + "grad_norm": 0.10363951325416565, + "learning_rate": 0.0001619714648197833, + "loss": 0.2399, + "step": 36570 + }, + { + "epoch": 0.5705996131528046, + "grad_norm": 1.3341947793960571, + "learning_rate": 0.0001619610657016285, + "loss": 0.1397, + "step": 36580 + }, + { + "epoch": 0.5707555999251264, + "grad_norm": 0.0668073296546936, + "learning_rate": 0.00016195066658347373, + "loss": 0.1091, + "step": 36590 + }, + { + "epoch": 0.570911586697448, + "grad_norm": 5.0486674308776855, + "learning_rate": 0.00016194026746531895, + "loss": 0.4893, + "step": 36600 + }, + { + "epoch": 0.5710675734697698, + "grad_norm": 1.2655692100524902, + "learning_rate": 0.00016192986834716417, + "loss": 0.3344, + "step": 36610 + }, + { + "epoch": 0.5712235602420914, + "grad_norm": 0.1458190381526947, + "learning_rate": 0.00016191946922900939, + "loss": 0.247, + "step": 36620 + }, + { + "epoch": 0.5713795470144132, + "grad_norm": 2.5789072513580322, + "learning_rate": 0.0001619090701108546, + "loss": 0.1546, + "step": 36630 + }, + { + "epoch": 0.5715355337867349, + "grad_norm": 1.191821575164795, + "learning_rate": 0.00016189867099269983, + "loss": 0.2215, + "step": 36640 + }, + { + "epoch": 0.5716915205590566, + "grad_norm": 8.163617134094238, + "learning_rate": 0.00016188827187454505, + "loss": 0.1696, + "step": 36650 + }, + { + "epoch": 0.5718475073313783, + "grad_norm": 3.162277936935425, + "learning_rate": 0.00016187787275639027, + "loss": 0.2906, + "step": 36660 + }, + { + "epoch": 0.5720034941037, + "grad_norm": 2.2368602752685547, + "learning_rate": 0.00016186747363823548, + "loss": 0.3806, + "step": 36670 + }, + { + "epoch": 0.5721594808760218, + "grad_norm": 2.0193803310394287, + "learning_rate": 0.0001618570745200807, + "loss": 0.1964, + "step": 36680 + }, + { + "epoch": 0.5723154676483434, + "grad_norm": 0.9131811261177063, + "learning_rate": 0.00016184667540192592, + "loss": 0.1902, + "step": 36690 + }, + { + "epoch": 0.5724714544206652, + "grad_norm": 0.21330870687961578, + "learning_rate": 0.00016183627628377114, + "loss": 0.2907, + "step": 36700 + }, + { + "epoch": 0.5726274411929868, + "grad_norm": 0.7145895957946777, + "learning_rate": 0.00016182587716561636, + "loss": 0.2574, + "step": 36710 + }, + { + "epoch": 0.5727834279653086, + "grad_norm": 1.0699433088302612, + "learning_rate": 0.00016181547804746158, + "loss": 0.1675, + "step": 36720 + }, + { + "epoch": 0.5729394147376302, + "grad_norm": 0.5066679120063782, + "learning_rate": 0.0001618050789293068, + "loss": 0.4487, + "step": 36730 + }, + { + "epoch": 0.573095401509952, + "grad_norm": 0.7482333183288574, + "learning_rate": 0.00016179467981115202, + "loss": 0.4101, + "step": 36740 + }, + { + "epoch": 0.5732513882822736, + "grad_norm": 0.15925532579421997, + "learning_rate": 0.00016178428069299724, + "loss": 0.2825, + "step": 36750 + }, + { + "epoch": 0.5734073750545954, + "grad_norm": 0.95689457654953, + "learning_rate": 0.00016177388157484246, + "loss": 0.2754, + "step": 36760 + }, + { + "epoch": 0.573563361826917, + "grad_norm": 2.15043568611145, + "learning_rate": 0.00016176348245668768, + "loss": 0.2073, + "step": 36770 + }, + { + "epoch": 0.5737193485992388, + "grad_norm": 2.651520252227783, + "learning_rate": 0.0001617530833385329, + "loss": 0.2517, + "step": 36780 + }, + { + "epoch": 0.5738753353715605, + "grad_norm": 2.1156702041625977, + "learning_rate": 0.00016174268422037812, + "loss": 0.3029, + "step": 36790 + }, + { + "epoch": 0.5740313221438822, + "grad_norm": 3.6200709342956543, + "learning_rate": 0.00016173228510222334, + "loss": 0.4174, + "step": 36800 + }, + { + "epoch": 0.574187308916204, + "grad_norm": 1.447937250137329, + "learning_rate": 0.00016172188598406856, + "loss": 0.1763, + "step": 36810 + }, + { + "epoch": 0.5743432956885256, + "grad_norm": 1.4020819664001465, + "learning_rate": 0.00016171148686591378, + "loss": 0.3994, + "step": 36820 + }, + { + "epoch": 0.5744992824608474, + "grad_norm": 1.4079807996749878, + "learning_rate": 0.000161701087747759, + "loss": 0.2051, + "step": 36830 + }, + { + "epoch": 0.574655269233169, + "grad_norm": 0.05268567427992821, + "learning_rate": 0.00016169068862960422, + "loss": 0.1773, + "step": 36840 + }, + { + "epoch": 0.5748112560054908, + "grad_norm": 2.248767614364624, + "learning_rate": 0.00016168028951144944, + "loss": 0.2534, + "step": 36850 + }, + { + "epoch": 0.5749672427778124, + "grad_norm": 0.3726302981376648, + "learning_rate": 0.00016166989039329466, + "loss": 0.328, + "step": 36860 + }, + { + "epoch": 0.5751232295501342, + "grad_norm": 1.0772110223770142, + "learning_rate": 0.00016165949127513988, + "loss": 0.2069, + "step": 36870 + }, + { + "epoch": 0.5752792163224558, + "grad_norm": 0.11339398473501205, + "learning_rate": 0.0001616490921569851, + "loss": 0.1827, + "step": 36880 + }, + { + "epoch": 0.5754352030947776, + "grad_norm": 1.2981103658676147, + "learning_rate": 0.00016163869303883032, + "loss": 0.1378, + "step": 36890 + }, + { + "epoch": 0.5755911898670992, + "grad_norm": 0.3302285075187683, + "learning_rate": 0.00016162829392067554, + "loss": 0.2311, + "step": 36900 + }, + { + "epoch": 0.575747176639421, + "grad_norm": 0.10168436169624329, + "learning_rate": 0.00016161789480252076, + "loss": 0.1678, + "step": 36910 + }, + { + "epoch": 0.5759031634117426, + "grad_norm": 1.6778357028961182, + "learning_rate": 0.00016160749568436598, + "loss": 0.2385, + "step": 36920 + }, + { + "epoch": 0.5760591501840644, + "grad_norm": 8.1710205078125, + "learning_rate": 0.0001615970965662112, + "loss": 0.3287, + "step": 36930 + }, + { + "epoch": 0.5762151369563862, + "grad_norm": 2.6481053829193115, + "learning_rate": 0.00016158669744805641, + "loss": 0.3119, + "step": 36940 + }, + { + "epoch": 0.5763711237287078, + "grad_norm": 0.3120105266571045, + "learning_rate": 0.00016157629832990163, + "loss": 0.1985, + "step": 36950 + }, + { + "epoch": 0.5765271105010296, + "grad_norm": 2.7221410274505615, + "learning_rate": 0.00016156589921174685, + "loss": 0.3697, + "step": 36960 + }, + { + "epoch": 0.5766830972733512, + "grad_norm": 0.34397706389427185, + "learning_rate": 0.00016155550009359207, + "loss": 0.2162, + "step": 36970 + }, + { + "epoch": 0.576839084045673, + "grad_norm": 4.6421003341674805, + "learning_rate": 0.0001615451009754373, + "loss": 0.2892, + "step": 36980 + }, + { + "epoch": 0.5769950708179946, + "grad_norm": 1.5395855903625488, + "learning_rate": 0.0001615347018572825, + "loss": 0.1258, + "step": 36990 + }, + { + "epoch": 0.5771510575903164, + "grad_norm": 0.4586782455444336, + "learning_rate": 0.00016152430273912773, + "loss": 0.1752, + "step": 37000 + }, + { + "epoch": 0.577307044362638, + "grad_norm": 0.4277658760547638, + "learning_rate": 0.00016151390362097295, + "loss": 0.1722, + "step": 37010 + }, + { + "epoch": 0.5774630311349598, + "grad_norm": 1.0282249450683594, + "learning_rate": 0.00016150350450281817, + "loss": 0.1361, + "step": 37020 + }, + { + "epoch": 0.5776190179072814, + "grad_norm": 0.2496921867132187, + "learning_rate": 0.0001614931053846634, + "loss": 0.0954, + "step": 37030 + }, + { + "epoch": 0.5777750046796032, + "grad_norm": 1.7900562286376953, + "learning_rate": 0.0001614827062665086, + "loss": 0.1284, + "step": 37040 + }, + { + "epoch": 0.5779309914519248, + "grad_norm": 0.6629091501235962, + "learning_rate": 0.00016147230714835383, + "loss": 0.2657, + "step": 37050 + }, + { + "epoch": 0.5780869782242466, + "grad_norm": 2.502882242202759, + "learning_rate": 0.00016146190803019905, + "loss": 0.207, + "step": 37060 + }, + { + "epoch": 0.5782429649965682, + "grad_norm": 1.486069917678833, + "learning_rate": 0.00016145150891204427, + "loss": 0.1232, + "step": 37070 + }, + { + "epoch": 0.57839895176889, + "grad_norm": 0.13543163239955902, + "learning_rate": 0.0001614411097938895, + "loss": 0.3245, + "step": 37080 + }, + { + "epoch": 0.5785549385412118, + "grad_norm": 3.9703898429870605, + "learning_rate": 0.0001614307106757347, + "loss": 0.3104, + "step": 37090 + }, + { + "epoch": 0.5787109253135334, + "grad_norm": 3.2236313819885254, + "learning_rate": 0.00016142031155757993, + "loss": 0.1739, + "step": 37100 + }, + { + "epoch": 0.5788669120858552, + "grad_norm": 0.3854866325855255, + "learning_rate": 0.00016140991243942515, + "loss": 0.084, + "step": 37110 + }, + { + "epoch": 0.5790228988581768, + "grad_norm": 0.1981169581413269, + "learning_rate": 0.00016139951332127037, + "loss": 0.2367, + "step": 37120 + }, + { + "epoch": 0.5791788856304986, + "grad_norm": 2.7315785884857178, + "learning_rate": 0.0001613891142031156, + "loss": 0.1816, + "step": 37130 + }, + { + "epoch": 0.5793348724028202, + "grad_norm": 0.48251378536224365, + "learning_rate": 0.0001613787150849608, + "loss": 0.2123, + "step": 37140 + }, + { + "epoch": 0.579490859175142, + "grad_norm": 5.0431413650512695, + "learning_rate": 0.00016136831596680603, + "loss": 0.2489, + "step": 37150 + }, + { + "epoch": 0.5796468459474636, + "grad_norm": 0.81981360912323, + "learning_rate": 0.00016135791684865125, + "loss": 0.5235, + "step": 37160 + }, + { + "epoch": 0.5798028327197854, + "grad_norm": 0.5701965689659119, + "learning_rate": 0.00016134751773049647, + "loss": 0.1709, + "step": 37170 + }, + { + "epoch": 0.579958819492107, + "grad_norm": 1.3905616998672485, + "learning_rate": 0.00016133711861234169, + "loss": 0.2641, + "step": 37180 + }, + { + "epoch": 0.5801148062644288, + "grad_norm": 2.281285047531128, + "learning_rate": 0.0001613267194941869, + "loss": 0.3915, + "step": 37190 + }, + { + "epoch": 0.5802707930367504, + "grad_norm": 1.2038309574127197, + "learning_rate": 0.00016131632037603213, + "loss": 0.0881, + "step": 37200 + }, + { + "epoch": 0.5804267798090722, + "grad_norm": 0.3096400201320648, + "learning_rate": 0.00016130592125787735, + "loss": 0.1341, + "step": 37210 + }, + { + "epoch": 0.5805827665813938, + "grad_norm": 0.10219905525445938, + "learning_rate": 0.00016129552213972256, + "loss": 0.0867, + "step": 37220 + }, + { + "epoch": 0.5807387533537156, + "grad_norm": 0.13869646191596985, + "learning_rate": 0.00016128512302156778, + "loss": 0.1621, + "step": 37230 + }, + { + "epoch": 0.5808947401260373, + "grad_norm": 0.7992938160896301, + "learning_rate": 0.000161274723903413, + "loss": 0.275, + "step": 37240 + }, + { + "epoch": 0.581050726898359, + "grad_norm": 0.2635972201824188, + "learning_rate": 0.00016126432478525822, + "loss": 0.2429, + "step": 37250 + }, + { + "epoch": 0.5812067136706808, + "grad_norm": 1.7211238145828247, + "learning_rate": 0.00016125392566710344, + "loss": 0.1541, + "step": 37260 + }, + { + "epoch": 0.5813627004430024, + "grad_norm": 2.2448506355285645, + "learning_rate": 0.00016124352654894866, + "loss": 0.3689, + "step": 37270 + }, + { + "epoch": 0.5815186872153242, + "grad_norm": 1.238708257675171, + "learning_rate": 0.00016123312743079388, + "loss": 0.3292, + "step": 37280 + }, + { + "epoch": 0.5816746739876458, + "grad_norm": 0.5670503377914429, + "learning_rate": 0.0001612227283126391, + "loss": 0.1151, + "step": 37290 + }, + { + "epoch": 0.5818306607599676, + "grad_norm": 2.287304401397705, + "learning_rate": 0.00016121232919448432, + "loss": 0.3558, + "step": 37300 + }, + { + "epoch": 0.5819866475322892, + "grad_norm": 0.8009629845619202, + "learning_rate": 0.00016120193007632954, + "loss": 0.1861, + "step": 37310 + }, + { + "epoch": 0.582142634304611, + "grad_norm": 1.1901466846466064, + "learning_rate": 0.00016119153095817476, + "loss": 0.089, + "step": 37320 + }, + { + "epoch": 0.5822986210769326, + "grad_norm": 1.7198141813278198, + "learning_rate": 0.00016118113184001998, + "loss": 0.2931, + "step": 37330 + }, + { + "epoch": 0.5824546078492544, + "grad_norm": 0.8832749724388123, + "learning_rate": 0.0001611707327218652, + "loss": 0.2342, + "step": 37340 + }, + { + "epoch": 0.582610594621576, + "grad_norm": 1.0623290538787842, + "learning_rate": 0.00016116033360371042, + "loss": 0.2317, + "step": 37350 + }, + { + "epoch": 0.5827665813938978, + "grad_norm": 0.5173699855804443, + "learning_rate": 0.00016114993448555564, + "loss": 0.1299, + "step": 37360 + }, + { + "epoch": 0.5829225681662195, + "grad_norm": 0.05118720978498459, + "learning_rate": 0.00016113953536740086, + "loss": 0.0786, + "step": 37370 + }, + { + "epoch": 0.5830785549385412, + "grad_norm": 2.0060107707977295, + "learning_rate": 0.00016112913624924608, + "loss": 0.266, + "step": 37380 + }, + { + "epoch": 0.5832345417108629, + "grad_norm": 5.121693134307861, + "learning_rate": 0.0001611187371310913, + "loss": 0.1824, + "step": 37390 + }, + { + "epoch": 0.5833905284831846, + "grad_norm": 0.3568089008331299, + "learning_rate": 0.00016110833801293652, + "loss": 0.3189, + "step": 37400 + }, + { + "epoch": 0.5835465152555064, + "grad_norm": 0.8631492853164673, + "learning_rate": 0.00016109793889478174, + "loss": 0.1724, + "step": 37410 + }, + { + "epoch": 0.583702502027828, + "grad_norm": 1.9214404821395874, + "learning_rate": 0.00016108753977662693, + "loss": 0.2094, + "step": 37420 + }, + { + "epoch": 0.5838584888001498, + "grad_norm": 1.9055646657943726, + "learning_rate": 0.00016107714065847218, + "loss": 0.1258, + "step": 37430 + }, + { + "epoch": 0.5840144755724714, + "grad_norm": 0.6614194512367249, + "learning_rate": 0.00016106674154031737, + "loss": 0.1966, + "step": 37440 + }, + { + "epoch": 0.5841704623447932, + "grad_norm": 0.286883682012558, + "learning_rate": 0.00016105634242216262, + "loss": 0.588, + "step": 37450 + }, + { + "epoch": 0.5843264491171148, + "grad_norm": 0.5599405765533447, + "learning_rate": 0.0001610459433040078, + "loss": 0.3414, + "step": 37460 + }, + { + "epoch": 0.5844824358894366, + "grad_norm": 1.5908915996551514, + "learning_rate": 0.00016103554418585306, + "loss": 0.1264, + "step": 37470 + }, + { + "epoch": 0.5846384226617583, + "grad_norm": 1.0729930400848389, + "learning_rate": 0.00016102514506769825, + "loss": 0.4815, + "step": 37480 + }, + { + "epoch": 0.58479440943408, + "grad_norm": 0.8912356495857239, + "learning_rate": 0.0001610147459495435, + "loss": 0.2092, + "step": 37490 + }, + { + "epoch": 0.5849503962064017, + "grad_norm": 0.24116197228431702, + "learning_rate": 0.0001610043468313887, + "loss": 0.1552, + "step": 37500 + }, + { + "epoch": 0.5851063829787234, + "grad_norm": 2.245192527770996, + "learning_rate": 0.00016099394771323393, + "loss": 0.0553, + "step": 37510 + }, + { + "epoch": 0.5852623697510451, + "grad_norm": 0.11538795381784439, + "learning_rate": 0.00016098354859507913, + "loss": 0.0889, + "step": 37520 + }, + { + "epoch": 0.5854183565233668, + "grad_norm": 1.0372689962387085, + "learning_rate": 0.00016097314947692437, + "loss": 0.3909, + "step": 37530 + }, + { + "epoch": 0.5855743432956885, + "grad_norm": 1.1946550607681274, + "learning_rate": 0.00016096275035876957, + "loss": 0.2167, + "step": 37540 + }, + { + "epoch": 0.5857303300680102, + "grad_norm": 1.3474591970443726, + "learning_rate": 0.0001609523512406148, + "loss": 0.1561, + "step": 37550 + }, + { + "epoch": 0.585886316840332, + "grad_norm": 0.008936616592109203, + "learning_rate": 0.00016094195212246, + "loss": 0.3991, + "step": 37560 + }, + { + "epoch": 0.5860423036126536, + "grad_norm": 1.8822320699691772, + "learning_rate": 0.00016093155300430525, + "loss": 0.2438, + "step": 37570 + }, + { + "epoch": 0.5861982903849754, + "grad_norm": 1.1707524061203003, + "learning_rate": 0.00016092115388615044, + "loss": 0.308, + "step": 37580 + }, + { + "epoch": 0.586354277157297, + "grad_norm": 1.679641842842102, + "learning_rate": 0.0001609107547679957, + "loss": 0.159, + "step": 37590 + }, + { + "epoch": 0.5865102639296188, + "grad_norm": 1.6885284185409546, + "learning_rate": 0.00016090035564984088, + "loss": 0.157, + "step": 37600 + }, + { + "epoch": 0.5866662507019405, + "grad_norm": 0.3876996636390686, + "learning_rate": 0.00016088995653168613, + "loss": 0.4067, + "step": 37610 + }, + { + "epoch": 0.5868222374742622, + "grad_norm": 1.387480616569519, + "learning_rate": 0.00016087955741353132, + "loss": 0.1898, + "step": 37620 + }, + { + "epoch": 0.5869782242465839, + "grad_norm": 0.03653861582279205, + "learning_rate": 0.00016086915829537657, + "loss": 0.1652, + "step": 37630 + }, + { + "epoch": 0.5871342110189056, + "grad_norm": 0.6258412003517151, + "learning_rate": 0.00016085875917722176, + "loss": 0.2558, + "step": 37640 + }, + { + "epoch": 0.5872901977912273, + "grad_norm": 0.10519164055585861, + "learning_rate": 0.000160848360059067, + "loss": 0.243, + "step": 37650 + }, + { + "epoch": 0.587446184563549, + "grad_norm": 0.7370131015777588, + "learning_rate": 0.0001608379609409122, + "loss": 0.2343, + "step": 37660 + }, + { + "epoch": 0.5876021713358707, + "grad_norm": 0.6884517073631287, + "learning_rate": 0.00016082756182275745, + "loss": 0.1847, + "step": 37670 + }, + { + "epoch": 0.5877581581081924, + "grad_norm": 0.6326120495796204, + "learning_rate": 0.00016081716270460264, + "loss": 0.2258, + "step": 37680 + }, + { + "epoch": 0.5879141448805141, + "grad_norm": 7.697889804840088, + "learning_rate": 0.0001608067635864479, + "loss": 0.2068, + "step": 37690 + }, + { + "epoch": 0.5880701316528358, + "grad_norm": 1.8753420114517212, + "learning_rate": 0.00016079636446829308, + "loss": 0.3087, + "step": 37700 + }, + { + "epoch": 0.5882261184251576, + "grad_norm": 0.8591625094413757, + "learning_rate": 0.00016078596535013833, + "loss": 0.1082, + "step": 37710 + }, + { + "epoch": 0.5883821051974792, + "grad_norm": 0.6124423146247864, + "learning_rate": 0.00016077556623198352, + "loss": 0.1536, + "step": 37720 + }, + { + "epoch": 0.588538091969801, + "grad_norm": 0.1400165557861328, + "learning_rate": 0.00016076516711382877, + "loss": 0.145, + "step": 37730 + }, + { + "epoch": 0.5886940787421227, + "grad_norm": 1.9879004955291748, + "learning_rate": 0.00016075476799567396, + "loss": 0.4986, + "step": 37740 + }, + { + "epoch": 0.5888500655144444, + "grad_norm": 0.08687327802181244, + "learning_rate": 0.0001607443688775192, + "loss": 0.3365, + "step": 37750 + }, + { + "epoch": 0.5890060522867661, + "grad_norm": 0.3442278206348419, + "learning_rate": 0.0001607339697593644, + "loss": 0.2274, + "step": 37760 + }, + { + "epoch": 0.5891620390590878, + "grad_norm": 0.06124915927648544, + "learning_rate": 0.00016072357064120965, + "loss": 0.1915, + "step": 37770 + }, + { + "epoch": 0.5893180258314095, + "grad_norm": 1.0996816158294678, + "learning_rate": 0.00016071317152305484, + "loss": 0.1573, + "step": 37780 + }, + { + "epoch": 0.5894740126037312, + "grad_norm": 0.14738436043262482, + "learning_rate": 0.00016070277240490008, + "loss": 0.2524, + "step": 37790 + }, + { + "epoch": 0.5896299993760529, + "grad_norm": 2.3905467987060547, + "learning_rate": 0.00016069237328674528, + "loss": 0.1838, + "step": 37800 + }, + { + "epoch": 0.5897859861483746, + "grad_norm": 2.0800178050994873, + "learning_rate": 0.00016068197416859052, + "loss": 0.3749, + "step": 37810 + }, + { + "epoch": 0.5899419729206963, + "grad_norm": 3.254204511642456, + "learning_rate": 0.00016067157505043574, + "loss": 0.3777, + "step": 37820 + }, + { + "epoch": 0.590097959693018, + "grad_norm": 2.8447892665863037, + "learning_rate": 0.00016066117593228096, + "loss": 0.5338, + "step": 37830 + }, + { + "epoch": 0.5902539464653397, + "grad_norm": 0.624733030796051, + "learning_rate": 0.00016065077681412618, + "loss": 0.2331, + "step": 37840 + }, + { + "epoch": 0.5904099332376614, + "grad_norm": 0.6863507628440857, + "learning_rate": 0.0001606403776959714, + "loss": 0.1747, + "step": 37850 + }, + { + "epoch": 0.5905659200099832, + "grad_norm": 0.6785943508148193, + "learning_rate": 0.00016062997857781662, + "loss": 0.2064, + "step": 37860 + }, + { + "epoch": 0.5907219067823049, + "grad_norm": 0.5739938616752625, + "learning_rate": 0.00016061957945966181, + "loss": 0.2368, + "step": 37870 + }, + { + "epoch": 0.5908778935546266, + "grad_norm": 0.0616387277841568, + "learning_rate": 0.00016060918034150706, + "loss": 0.2224, + "step": 37880 + }, + { + "epoch": 0.5910338803269483, + "grad_norm": 2.745957136154175, + "learning_rate": 0.00016059878122335225, + "loss": 0.3161, + "step": 37890 + }, + { + "epoch": 0.59118986709927, + "grad_norm": 2.52472186088562, + "learning_rate": 0.0001605883821051975, + "loss": 0.1713, + "step": 37900 + }, + { + "epoch": 0.5913458538715917, + "grad_norm": 1.5691877603530884, + "learning_rate": 0.0001605779829870427, + "loss": 0.1402, + "step": 37910 + }, + { + "epoch": 0.5915018406439134, + "grad_norm": 1.848261833190918, + "learning_rate": 0.00016056758386888794, + "loss": 0.2305, + "step": 37920 + }, + { + "epoch": 0.5916578274162351, + "grad_norm": 0.36622482538223267, + "learning_rate": 0.00016055718475073313, + "loss": 0.1834, + "step": 37930 + }, + { + "epoch": 0.5918138141885568, + "grad_norm": 0.21201446652412415, + "learning_rate": 0.00016054678563257838, + "loss": 0.4682, + "step": 37940 + }, + { + "epoch": 0.5919698009608785, + "grad_norm": 0.38921502232551575, + "learning_rate": 0.00016053638651442357, + "loss": 0.2892, + "step": 37950 + }, + { + "epoch": 0.5921257877332002, + "grad_norm": 0.20777840912342072, + "learning_rate": 0.00016052598739626882, + "loss": 0.104, + "step": 37960 + }, + { + "epoch": 0.5922817745055219, + "grad_norm": 0.7090150713920593, + "learning_rate": 0.000160515588278114, + "loss": 0.2592, + "step": 37970 + }, + { + "epoch": 0.5924377612778436, + "grad_norm": 1.9939709901809692, + "learning_rate": 0.00016050518915995926, + "loss": 0.247, + "step": 37980 + }, + { + "epoch": 0.5925937480501653, + "grad_norm": 0.8696603178977966, + "learning_rate": 0.00016049479004180445, + "loss": 0.301, + "step": 37990 + }, + { + "epoch": 0.592749734822487, + "grad_norm": 0.4119095206260681, + "learning_rate": 0.0001604843909236497, + "loss": 0.2011, + "step": 38000 + }, + { + "epoch": 0.5929057215948088, + "grad_norm": 0.060975607484579086, + "learning_rate": 0.0001604739918054949, + "loss": 0.2286, + "step": 38010 + }, + { + "epoch": 0.5930617083671305, + "grad_norm": 0.7039875388145447, + "learning_rate": 0.00016046359268734014, + "loss": 0.0999, + "step": 38020 + }, + { + "epoch": 0.5932176951394522, + "grad_norm": 0.28402045369148254, + "learning_rate": 0.00016045319356918533, + "loss": 0.1835, + "step": 38030 + }, + { + "epoch": 0.5933736819117739, + "grad_norm": 2.217988967895508, + "learning_rate": 0.00016044279445103058, + "loss": 0.0785, + "step": 38040 + }, + { + "epoch": 0.5935296686840956, + "grad_norm": 1.4856069087982178, + "learning_rate": 0.00016043239533287577, + "loss": 0.2528, + "step": 38050 + }, + { + "epoch": 0.5936856554564173, + "grad_norm": 0.9875519871711731, + "learning_rate": 0.00016042199621472101, + "loss": 0.4445, + "step": 38060 + }, + { + "epoch": 0.593841642228739, + "grad_norm": 2.773216724395752, + "learning_rate": 0.0001604115970965662, + "loss": 0.5643, + "step": 38070 + }, + { + "epoch": 0.5939976290010607, + "grad_norm": 1.7584549188613892, + "learning_rate": 0.00016040119797841145, + "loss": 0.1446, + "step": 38080 + }, + { + "epoch": 0.5941536157733824, + "grad_norm": 1.2456908226013184, + "learning_rate": 0.00016039079886025665, + "loss": 0.1896, + "step": 38090 + }, + { + "epoch": 0.5943096025457041, + "grad_norm": 0.021098516881465912, + "learning_rate": 0.0001603803997421019, + "loss": 0.0876, + "step": 38100 + }, + { + "epoch": 0.5944655893180258, + "grad_norm": 1.8131749629974365, + "learning_rate": 0.00016037000062394709, + "loss": 0.1639, + "step": 38110 + }, + { + "epoch": 0.5946215760903475, + "grad_norm": 0.41819462180137634, + "learning_rate": 0.00016035960150579233, + "loss": 0.1311, + "step": 38120 + }, + { + "epoch": 0.5947775628626693, + "grad_norm": 0.6691219806671143, + "learning_rate": 0.00016034920238763753, + "loss": 0.1625, + "step": 38130 + }, + { + "epoch": 0.5949335496349909, + "grad_norm": 0.07854852080345154, + "learning_rate": 0.00016033880326948277, + "loss": 0.3177, + "step": 38140 + }, + { + "epoch": 0.5950895364073127, + "grad_norm": 0.7819689512252808, + "learning_rate": 0.00016032840415132796, + "loss": 0.4225, + "step": 38150 + }, + { + "epoch": 0.5952455231796344, + "grad_norm": 5.3962836265563965, + "learning_rate": 0.0001603180050331732, + "loss": 0.123, + "step": 38160 + }, + { + "epoch": 0.5954015099519561, + "grad_norm": 0.7460303902626038, + "learning_rate": 0.0001603076059150184, + "loss": 0.2182, + "step": 38170 + }, + { + "epoch": 0.5955574967242778, + "grad_norm": 0.26433685421943665, + "learning_rate": 0.00016029720679686365, + "loss": 0.1894, + "step": 38180 + }, + { + "epoch": 0.5957134834965995, + "grad_norm": 0.8701320290565491, + "learning_rate": 0.00016028680767870884, + "loss": 0.2137, + "step": 38190 + }, + { + "epoch": 0.5958694702689212, + "grad_norm": 3.9774930477142334, + "learning_rate": 0.0001602764085605541, + "loss": 0.4637, + "step": 38200 + }, + { + "epoch": 0.5960254570412429, + "grad_norm": 0.5801145434379578, + "learning_rate": 0.00016026600944239928, + "loss": 0.3036, + "step": 38210 + }, + { + "epoch": 0.5961814438135646, + "grad_norm": 1.1890723705291748, + "learning_rate": 0.00016025561032424453, + "loss": 0.0988, + "step": 38220 + }, + { + "epoch": 0.5963374305858863, + "grad_norm": 0.18521073460578918, + "learning_rate": 0.00016024521120608972, + "loss": 0.1529, + "step": 38230 + }, + { + "epoch": 0.596493417358208, + "grad_norm": 2.2602827548980713, + "learning_rate": 0.00016023481208793497, + "loss": 0.2815, + "step": 38240 + }, + { + "epoch": 0.5966494041305297, + "grad_norm": 0.4060989022254944, + "learning_rate": 0.00016022441296978016, + "loss": 0.3466, + "step": 38250 + }, + { + "epoch": 0.5968053909028515, + "grad_norm": 1.4743244647979736, + "learning_rate": 0.0001602140138516254, + "loss": 0.3561, + "step": 38260 + }, + { + "epoch": 0.5969613776751731, + "grad_norm": 5.080365180969238, + "learning_rate": 0.0001602036147334706, + "loss": 0.2914, + "step": 38270 + }, + { + "epoch": 0.5971173644474949, + "grad_norm": 0.16149812936782837, + "learning_rate": 0.00016019321561531585, + "loss": 0.1871, + "step": 38280 + }, + { + "epoch": 0.5972733512198165, + "grad_norm": 1.7535364627838135, + "learning_rate": 0.00016018281649716104, + "loss": 0.4905, + "step": 38290 + }, + { + "epoch": 0.5974293379921383, + "grad_norm": 0.4103987216949463, + "learning_rate": 0.00016017241737900629, + "loss": 0.3813, + "step": 38300 + }, + { + "epoch": 0.59758532476446, + "grad_norm": 0.5315369367599487, + "learning_rate": 0.00016016201826085148, + "loss": 0.3897, + "step": 38310 + }, + { + "epoch": 0.5977413115367817, + "grad_norm": 3.4553463459014893, + "learning_rate": 0.0001601516191426967, + "loss": 0.2047, + "step": 38320 + }, + { + "epoch": 0.5978972983091034, + "grad_norm": 2.225468397140503, + "learning_rate": 0.00016014122002454192, + "loss": 0.4028, + "step": 38330 + }, + { + "epoch": 0.5980532850814251, + "grad_norm": 2.000885486602783, + "learning_rate": 0.00016013082090638714, + "loss": 0.391, + "step": 38340 + }, + { + "epoch": 0.5982092718537468, + "grad_norm": 0.7886672616004944, + "learning_rate": 0.00016012042178823236, + "loss": 0.2994, + "step": 38350 + }, + { + "epoch": 0.5983652586260685, + "grad_norm": 0.5279917120933533, + "learning_rate": 0.00016011002267007758, + "loss": 0.1855, + "step": 38360 + }, + { + "epoch": 0.5985212453983902, + "grad_norm": 0.9789942502975464, + "learning_rate": 0.0001600996235519228, + "loss": 0.2122, + "step": 38370 + }, + { + "epoch": 0.5986772321707119, + "grad_norm": 0.6289750933647156, + "learning_rate": 0.00016008922443376802, + "loss": 0.3737, + "step": 38380 + }, + { + "epoch": 0.5988332189430337, + "grad_norm": 1.3074733018875122, + "learning_rate": 0.00016007882531561324, + "loss": 0.2382, + "step": 38390 + }, + { + "epoch": 0.5989892057153553, + "grad_norm": 1.745284914970398, + "learning_rate": 0.00016006842619745846, + "loss": 0.3167, + "step": 38400 + }, + { + "epoch": 0.5991451924876771, + "grad_norm": 0.1694452464580536, + "learning_rate": 0.00016005802707930368, + "loss": 0.1632, + "step": 38410 + }, + { + "epoch": 0.5993011792599987, + "grad_norm": 1.55560302734375, + "learning_rate": 0.0001600476279611489, + "loss": 0.2371, + "step": 38420 + }, + { + "epoch": 0.5994571660323205, + "grad_norm": 1.044997215270996, + "learning_rate": 0.00016003722884299411, + "loss": 0.0896, + "step": 38430 + }, + { + "epoch": 0.5996131528046421, + "grad_norm": 0.25250110030174255, + "learning_rate": 0.00016002682972483933, + "loss": 0.3096, + "step": 38440 + }, + { + "epoch": 0.5997691395769639, + "grad_norm": 0.1631859689950943, + "learning_rate": 0.00016001643060668455, + "loss": 0.2004, + "step": 38450 + }, + { + "epoch": 0.5999251263492856, + "grad_norm": 0.34967318177223206, + "learning_rate": 0.00016000603148852977, + "loss": 0.1476, + "step": 38460 + }, + { + "epoch": 0.6000811131216073, + "grad_norm": 1.8604000806808472, + "learning_rate": 0.000159995632370375, + "loss": 0.1884, + "step": 38470 + }, + { + "epoch": 0.600237099893929, + "grad_norm": 1.626848578453064, + "learning_rate": 0.0001599852332522202, + "loss": 0.5452, + "step": 38480 + }, + { + "epoch": 0.6003930866662507, + "grad_norm": 0.41935989260673523, + "learning_rate": 0.00015997483413406543, + "loss": 0.3047, + "step": 38490 + }, + { + "epoch": 0.6005490734385724, + "grad_norm": 1.355086088180542, + "learning_rate": 0.00015996443501591065, + "loss": 0.1246, + "step": 38500 + }, + { + "epoch": 0.6007050602108941, + "grad_norm": 1.5382829904556274, + "learning_rate": 0.00015995403589775587, + "loss": 0.2324, + "step": 38510 + }, + { + "epoch": 0.6008610469832159, + "grad_norm": 1.0250797271728516, + "learning_rate": 0.0001599436367796011, + "loss": 0.181, + "step": 38520 + }, + { + "epoch": 0.6010170337555375, + "grad_norm": 0.07823364436626434, + "learning_rate": 0.0001599332376614463, + "loss": 0.1492, + "step": 38530 + }, + { + "epoch": 0.6011730205278593, + "grad_norm": 0.07573071867227554, + "learning_rate": 0.00015992283854329153, + "loss": 0.1256, + "step": 38540 + }, + { + "epoch": 0.6013290073001809, + "grad_norm": 0.5608569383621216, + "learning_rate": 0.00015991243942513675, + "loss": 0.378, + "step": 38550 + }, + { + "epoch": 0.6014849940725027, + "grad_norm": 0.7001075744628906, + "learning_rate": 0.00015990204030698197, + "loss": 0.2518, + "step": 38560 + }, + { + "epoch": 0.6016409808448243, + "grad_norm": 4.868145942687988, + "learning_rate": 0.0001598916411888272, + "loss": 0.1857, + "step": 38570 + }, + { + "epoch": 0.6017969676171461, + "grad_norm": 0.05653705075383186, + "learning_rate": 0.0001598812420706724, + "loss": 0.1918, + "step": 38580 + }, + { + "epoch": 0.6019529543894677, + "grad_norm": 0.8556832671165466, + "learning_rate": 0.00015987084295251763, + "loss": 0.1926, + "step": 38590 + }, + { + "epoch": 0.6021089411617895, + "grad_norm": 1.1988335847854614, + "learning_rate": 0.00015986044383436285, + "loss": 0.3884, + "step": 38600 + }, + { + "epoch": 0.6022649279341112, + "grad_norm": 1.5978832244873047, + "learning_rate": 0.00015985004471620807, + "loss": 0.1986, + "step": 38610 + }, + { + "epoch": 0.6024209147064329, + "grad_norm": 1.084174633026123, + "learning_rate": 0.0001598396455980533, + "loss": 0.4602, + "step": 38620 + }, + { + "epoch": 0.6025769014787546, + "grad_norm": 1.5116629600524902, + "learning_rate": 0.0001598292464798985, + "loss": 0.292, + "step": 38630 + }, + { + "epoch": 0.6027328882510763, + "grad_norm": 0.23945322632789612, + "learning_rate": 0.00015981884736174373, + "loss": 0.1904, + "step": 38640 + }, + { + "epoch": 0.602888875023398, + "grad_norm": 0.348156601190567, + "learning_rate": 0.00015980844824358895, + "loss": 0.1442, + "step": 38650 + }, + { + "epoch": 0.6030448617957197, + "grad_norm": 1.876736044883728, + "learning_rate": 0.00015979804912543417, + "loss": 0.248, + "step": 38660 + }, + { + "epoch": 0.6032008485680415, + "grad_norm": 0.2564373016357422, + "learning_rate": 0.00015978765000727939, + "loss": 0.2429, + "step": 38670 + }, + { + "epoch": 0.6033568353403631, + "grad_norm": 1.824837327003479, + "learning_rate": 0.0001597772508891246, + "loss": 0.2593, + "step": 38680 + }, + { + "epoch": 0.6035128221126849, + "grad_norm": 1.6967028379440308, + "learning_rate": 0.00015976685177096985, + "loss": 0.3918, + "step": 38690 + }, + { + "epoch": 0.6036688088850065, + "grad_norm": 1.6629718542099, + "learning_rate": 0.00015975645265281504, + "loss": 0.2529, + "step": 38700 + }, + { + "epoch": 0.6038247956573283, + "grad_norm": 0.19047370553016663, + "learning_rate": 0.0001597460535346603, + "loss": 0.4545, + "step": 38710 + }, + { + "epoch": 0.6039807824296499, + "grad_norm": 0.6910511255264282, + "learning_rate": 0.00015973565441650548, + "loss": 0.2226, + "step": 38720 + }, + { + "epoch": 0.6041367692019717, + "grad_norm": 1.9387328624725342, + "learning_rate": 0.00015972525529835073, + "loss": 0.5561, + "step": 38730 + }, + { + "epoch": 0.6042927559742933, + "grad_norm": 1.4491708278656006, + "learning_rate": 0.00015971485618019592, + "loss": 0.2987, + "step": 38740 + }, + { + "epoch": 0.6044487427466151, + "grad_norm": 0.32534459233283997, + "learning_rate": 0.00015970445706204117, + "loss": 0.1683, + "step": 38750 + }, + { + "epoch": 0.6046047295189368, + "grad_norm": 0.17335011065006256, + "learning_rate": 0.00015969405794388636, + "loss": 0.1554, + "step": 38760 + }, + { + "epoch": 0.6047607162912585, + "grad_norm": 0.39481663703918457, + "learning_rate": 0.00015968365882573158, + "loss": 0.186, + "step": 38770 + }, + { + "epoch": 0.6049167030635803, + "grad_norm": 0.10422719269990921, + "learning_rate": 0.0001596732597075768, + "loss": 0.1599, + "step": 38780 + }, + { + "epoch": 0.6050726898359019, + "grad_norm": 2.288064479827881, + "learning_rate": 0.00015966286058942202, + "loss": 0.1944, + "step": 38790 + }, + { + "epoch": 0.6052286766082237, + "grad_norm": 1.0062401294708252, + "learning_rate": 0.00015965246147126724, + "loss": 0.176, + "step": 38800 + }, + { + "epoch": 0.6053846633805453, + "grad_norm": 0.21470843255519867, + "learning_rate": 0.00015964206235311246, + "loss": 0.183, + "step": 38810 + }, + { + "epoch": 0.6055406501528671, + "grad_norm": 2.1730270385742188, + "learning_rate": 0.00015963166323495768, + "loss": 0.1788, + "step": 38820 + }, + { + "epoch": 0.6056966369251887, + "grad_norm": 0.6188749074935913, + "learning_rate": 0.0001596212641168029, + "loss": 0.2746, + "step": 38830 + }, + { + "epoch": 0.6058526236975105, + "grad_norm": 1.283058524131775, + "learning_rate": 0.00015961086499864812, + "loss": 0.1066, + "step": 38840 + }, + { + "epoch": 0.6060086104698321, + "grad_norm": 1.1469789743423462, + "learning_rate": 0.00015960046588049334, + "loss": 0.0692, + "step": 38850 + }, + { + "epoch": 0.6061645972421539, + "grad_norm": 1.9450418949127197, + "learning_rate": 0.00015959006676233856, + "loss": 0.2154, + "step": 38860 + }, + { + "epoch": 0.6063205840144755, + "grad_norm": 7.193401336669922, + "learning_rate": 0.00015957966764418378, + "loss": 0.218, + "step": 38870 + }, + { + "epoch": 0.6064765707867973, + "grad_norm": 3.5315003395080566, + "learning_rate": 0.000159569268526029, + "loss": 0.312, + "step": 38880 + }, + { + "epoch": 0.6066325575591189, + "grad_norm": 1.2873650789260864, + "learning_rate": 0.00015955886940787422, + "loss": 0.2587, + "step": 38890 + }, + { + "epoch": 0.6067885443314407, + "grad_norm": 1.1193941831588745, + "learning_rate": 0.00015954847028971944, + "loss": 0.2358, + "step": 38900 + }, + { + "epoch": 0.6069445311037625, + "grad_norm": 1.0471165180206299, + "learning_rate": 0.00015953807117156466, + "loss": 0.2639, + "step": 38910 + }, + { + "epoch": 0.6071005178760841, + "grad_norm": 1.7760034799575806, + "learning_rate": 0.00015952767205340988, + "loss": 0.248, + "step": 38920 + }, + { + "epoch": 0.6072565046484059, + "grad_norm": 0.050742994993925095, + "learning_rate": 0.0001595172729352551, + "loss": 0.1287, + "step": 38930 + }, + { + "epoch": 0.6074124914207275, + "grad_norm": 0.03016967698931694, + "learning_rate": 0.00015950687381710032, + "loss": 0.1843, + "step": 38940 + }, + { + "epoch": 0.6075684781930493, + "grad_norm": 0.6029847860336304, + "learning_rate": 0.00015949647469894554, + "loss": 0.2294, + "step": 38950 + }, + { + "epoch": 0.6077244649653709, + "grad_norm": 0.2955610454082489, + "learning_rate": 0.00015948607558079076, + "loss": 0.3336, + "step": 38960 + }, + { + "epoch": 0.6078804517376927, + "grad_norm": 1.9104161262512207, + "learning_rate": 0.00015947567646263597, + "loss": 0.189, + "step": 38970 + }, + { + "epoch": 0.6080364385100143, + "grad_norm": 0.09197711199522018, + "learning_rate": 0.0001594652773444812, + "loss": 0.1612, + "step": 38980 + }, + { + "epoch": 0.6081924252823361, + "grad_norm": 1.4233628511428833, + "learning_rate": 0.00015945487822632641, + "loss": 0.1521, + "step": 38990 + }, + { + "epoch": 0.6083484120546577, + "grad_norm": 9.787493705749512, + "learning_rate": 0.00015944447910817163, + "loss": 0.135, + "step": 39000 + }, + { + "epoch": 0.6085043988269795, + "grad_norm": 1.3848742246627808, + "learning_rate": 0.00015943407999001685, + "loss": 0.2268, + "step": 39010 + }, + { + "epoch": 0.6086603855993011, + "grad_norm": 0.1524626612663269, + "learning_rate": 0.00015942368087186207, + "loss": 0.3154, + "step": 39020 + }, + { + "epoch": 0.6088163723716229, + "grad_norm": 2.1885435581207275, + "learning_rate": 0.0001594132817537073, + "loss": 0.3174, + "step": 39030 + }, + { + "epoch": 0.6089723591439445, + "grad_norm": 1.4719898700714111, + "learning_rate": 0.0001594028826355525, + "loss": 0.2959, + "step": 39040 + }, + { + "epoch": 0.6091283459162663, + "grad_norm": 0.7844040989875793, + "learning_rate": 0.00015939248351739773, + "loss": 0.3021, + "step": 39050 + }, + { + "epoch": 0.6092843326885881, + "grad_norm": 0.17660890519618988, + "learning_rate": 0.00015938208439924295, + "loss": 0.3608, + "step": 39060 + }, + { + "epoch": 0.6094403194609097, + "grad_norm": 0.24621741473674774, + "learning_rate": 0.00015937168528108817, + "loss": 0.1461, + "step": 39070 + }, + { + "epoch": 0.6095963062332315, + "grad_norm": 0.7903050184249878, + "learning_rate": 0.0001593612861629334, + "loss": 0.1195, + "step": 39080 + }, + { + "epoch": 0.6097522930055531, + "grad_norm": 0.04043950140476227, + "learning_rate": 0.0001593508870447786, + "loss": 0.1938, + "step": 39090 + }, + { + "epoch": 0.6099082797778749, + "grad_norm": 0.7602111101150513, + "learning_rate": 0.00015934048792662383, + "loss": 0.1732, + "step": 39100 + }, + { + "epoch": 0.6100642665501965, + "grad_norm": 0.7902756333351135, + "learning_rate": 0.00015933008880846905, + "loss": 0.2351, + "step": 39110 + }, + { + "epoch": 0.6102202533225183, + "grad_norm": 1.7035608291625977, + "learning_rate": 0.00015931968969031427, + "loss": 0.2409, + "step": 39120 + }, + { + "epoch": 0.6103762400948399, + "grad_norm": 0.7998217940330505, + "learning_rate": 0.0001593092905721595, + "loss": 0.3862, + "step": 39130 + }, + { + "epoch": 0.6105322268671617, + "grad_norm": 0.3530103862285614, + "learning_rate": 0.0001592988914540047, + "loss": 0.2176, + "step": 39140 + }, + { + "epoch": 0.6106882136394833, + "grad_norm": 0.9218409657478333, + "learning_rate": 0.00015928849233584993, + "loss": 0.1362, + "step": 39150 + }, + { + "epoch": 0.6108442004118051, + "grad_norm": 0.643829345703125, + "learning_rate": 0.00015927809321769515, + "loss": 0.2309, + "step": 39160 + }, + { + "epoch": 0.6110001871841267, + "grad_norm": 0.1810176521539688, + "learning_rate": 0.00015926769409954037, + "loss": 0.2461, + "step": 39170 + }, + { + "epoch": 0.6111561739564485, + "grad_norm": 1.3435649871826172, + "learning_rate": 0.0001592572949813856, + "loss": 0.2937, + "step": 39180 + }, + { + "epoch": 0.6113121607287701, + "grad_norm": 1.1048580408096313, + "learning_rate": 0.0001592468958632308, + "loss": 0.2983, + "step": 39190 + }, + { + "epoch": 0.6114681475010919, + "grad_norm": 0.09463644027709961, + "learning_rate": 0.00015923649674507603, + "loss": 0.2318, + "step": 39200 + }, + { + "epoch": 0.6116241342734137, + "grad_norm": 0.9329594969749451, + "learning_rate": 0.00015922609762692125, + "loss": 0.3095, + "step": 39210 + }, + { + "epoch": 0.6117801210457353, + "grad_norm": 1.8576663732528687, + "learning_rate": 0.00015921569850876647, + "loss": 0.2336, + "step": 39220 + }, + { + "epoch": 0.6119361078180571, + "grad_norm": 1.6159908771514893, + "learning_rate": 0.00015920529939061169, + "loss": 0.1745, + "step": 39230 + }, + { + "epoch": 0.6120920945903787, + "grad_norm": 1.289712905883789, + "learning_rate": 0.0001591949002724569, + "loss": 0.2031, + "step": 39240 + }, + { + "epoch": 0.6122480813627005, + "grad_norm": 0.4183761179447174, + "learning_rate": 0.00015918450115430212, + "loss": 0.2911, + "step": 39250 + }, + { + "epoch": 0.6124040681350221, + "grad_norm": 2.125042200088501, + "learning_rate": 0.00015917410203614734, + "loss": 0.1086, + "step": 39260 + }, + { + "epoch": 0.6125600549073439, + "grad_norm": 3.527683973312378, + "learning_rate": 0.00015916370291799256, + "loss": 0.3356, + "step": 39270 + }, + { + "epoch": 0.6127160416796655, + "grad_norm": 1.1525741815567017, + "learning_rate": 0.00015915330379983778, + "loss": 0.222, + "step": 39280 + }, + { + "epoch": 0.6128720284519873, + "grad_norm": 0.5210650563240051, + "learning_rate": 0.000159142904681683, + "loss": 0.1563, + "step": 39290 + }, + { + "epoch": 0.6130280152243089, + "grad_norm": 0.2881133258342743, + "learning_rate": 0.00015913250556352822, + "loss": 0.346, + "step": 39300 + }, + { + "epoch": 0.6131840019966307, + "grad_norm": 1.7034498453140259, + "learning_rate": 0.00015912210644537344, + "loss": 0.1822, + "step": 39310 + }, + { + "epoch": 0.6133399887689523, + "grad_norm": 2.6305196285247803, + "learning_rate": 0.00015911170732721866, + "loss": 0.3373, + "step": 39320 + }, + { + "epoch": 0.6134959755412741, + "grad_norm": 0.4541589915752411, + "learning_rate": 0.00015910130820906388, + "loss": 0.2146, + "step": 39330 + }, + { + "epoch": 0.6136519623135958, + "grad_norm": 0.03301222622394562, + "learning_rate": 0.0001590909090909091, + "loss": 0.1787, + "step": 39340 + }, + { + "epoch": 0.6138079490859175, + "grad_norm": 0.7106437087059021, + "learning_rate": 0.00015908050997275432, + "loss": 0.2315, + "step": 39350 + }, + { + "epoch": 0.6139639358582393, + "grad_norm": 0.381320059299469, + "learning_rate": 0.00015907011085459954, + "loss": 0.1917, + "step": 39360 + }, + { + "epoch": 0.6141199226305609, + "grad_norm": 0.5418739914894104, + "learning_rate": 0.00015905971173644476, + "loss": 0.1571, + "step": 39370 + }, + { + "epoch": 0.6142759094028827, + "grad_norm": 0.3135250210762024, + "learning_rate": 0.00015904931261828998, + "loss": 0.0858, + "step": 39380 + }, + { + "epoch": 0.6144318961752043, + "grad_norm": 1.6296218633651733, + "learning_rate": 0.0001590389135001352, + "loss": 0.2297, + "step": 39390 + }, + { + "epoch": 0.6145878829475261, + "grad_norm": 0.5768391489982605, + "learning_rate": 0.00015902851438198042, + "loss": 0.1699, + "step": 39400 + }, + { + "epoch": 0.6147438697198477, + "grad_norm": 25.794370651245117, + "learning_rate": 0.00015901811526382564, + "loss": 0.1903, + "step": 39410 + }, + { + "epoch": 0.6148998564921695, + "grad_norm": 0.46698322892189026, + "learning_rate": 0.00015900771614567086, + "loss": 0.1415, + "step": 39420 + }, + { + "epoch": 0.6150558432644911, + "grad_norm": 1.3042336702346802, + "learning_rate": 0.00015899731702751608, + "loss": 0.1604, + "step": 39430 + }, + { + "epoch": 0.6152118300368129, + "grad_norm": 1.2311556339263916, + "learning_rate": 0.0001589869179093613, + "loss": 0.2555, + "step": 39440 + }, + { + "epoch": 0.6153678168091345, + "grad_norm": 0.6290935277938843, + "learning_rate": 0.00015897651879120652, + "loss": 0.1723, + "step": 39450 + }, + { + "epoch": 0.6155238035814563, + "grad_norm": 1.4786419868469238, + "learning_rate": 0.00015896611967305174, + "loss": 0.3717, + "step": 39460 + }, + { + "epoch": 0.615679790353778, + "grad_norm": 0.7589457631111145, + "learning_rate": 0.00015895572055489696, + "loss": 0.1912, + "step": 39470 + }, + { + "epoch": 0.6158357771260997, + "grad_norm": 0.10508652776479721, + "learning_rate": 0.00015894532143674218, + "loss": 0.1353, + "step": 39480 + }, + { + "epoch": 0.6159917638984214, + "grad_norm": 0.07394446432590485, + "learning_rate": 0.0001589349223185874, + "loss": 0.1195, + "step": 39490 + }, + { + "epoch": 0.6161477506707431, + "grad_norm": 0.11420007050037384, + "learning_rate": 0.00015892452320043262, + "loss": 0.0976, + "step": 39500 + }, + { + "epoch": 0.6163037374430649, + "grad_norm": 1.2442359924316406, + "learning_rate": 0.00015891412408227784, + "loss": 0.1084, + "step": 39510 + }, + { + "epoch": 0.6164597242153865, + "grad_norm": 0.4080588221549988, + "learning_rate": 0.00015890372496412306, + "loss": 0.2191, + "step": 39520 + }, + { + "epoch": 0.6166157109877083, + "grad_norm": 0.25214430689811707, + "learning_rate": 0.00015889332584596827, + "loss": 0.5548, + "step": 39530 + }, + { + "epoch": 0.6167716977600299, + "grad_norm": 0.41108590364456177, + "learning_rate": 0.0001588829267278135, + "loss": 0.4213, + "step": 39540 + }, + { + "epoch": 0.6169276845323517, + "grad_norm": 3.3010945320129395, + "learning_rate": 0.00015887252760965871, + "loss": 0.3946, + "step": 39550 + }, + { + "epoch": 0.6170836713046733, + "grad_norm": 0.25975245237350464, + "learning_rate": 0.00015886212849150393, + "loss": 0.3045, + "step": 39560 + }, + { + "epoch": 0.6172396580769951, + "grad_norm": 1.255585789680481, + "learning_rate": 0.00015885172937334915, + "loss": 0.2249, + "step": 39570 + }, + { + "epoch": 0.6173956448493167, + "grad_norm": 1.9135524034500122, + "learning_rate": 0.00015884133025519437, + "loss": 0.2401, + "step": 39580 + }, + { + "epoch": 0.6175516316216385, + "grad_norm": 2.1866586208343506, + "learning_rate": 0.0001588309311370396, + "loss": 0.1726, + "step": 39590 + }, + { + "epoch": 0.6177076183939602, + "grad_norm": 2.487704277038574, + "learning_rate": 0.0001588205320188848, + "loss": 0.2728, + "step": 39600 + }, + { + "epoch": 0.6178636051662819, + "grad_norm": 0.7800028920173645, + "learning_rate": 0.00015881013290073003, + "loss": 0.1522, + "step": 39610 + }, + { + "epoch": 0.6180195919386036, + "grad_norm": 3.468012571334839, + "learning_rate": 0.00015879973378257525, + "loss": 0.1984, + "step": 39620 + }, + { + "epoch": 0.6181755787109253, + "grad_norm": 1.5809861421585083, + "learning_rate": 0.00015878933466442047, + "loss": 0.2632, + "step": 39630 + }, + { + "epoch": 0.618331565483247, + "grad_norm": 0.5371276140213013, + "learning_rate": 0.0001587789355462657, + "loss": 0.1915, + "step": 39640 + }, + { + "epoch": 0.6184875522555687, + "grad_norm": 4.711477279663086, + "learning_rate": 0.0001587685364281109, + "loss": 0.5271, + "step": 39650 + }, + { + "epoch": 0.6186435390278905, + "grad_norm": 1.5581507682800293, + "learning_rate": 0.00015875813730995613, + "loss": 0.2482, + "step": 39660 + }, + { + "epoch": 0.6187995258002121, + "grad_norm": 1.7416914701461792, + "learning_rate": 0.00015874773819180132, + "loss": 0.2018, + "step": 39670 + }, + { + "epoch": 0.6189555125725339, + "grad_norm": 0.6761413216590881, + "learning_rate": 0.00015873733907364657, + "loss": 0.1048, + "step": 39680 + }, + { + "epoch": 0.6191114993448555, + "grad_norm": 1.2344286441802979, + "learning_rate": 0.00015872693995549176, + "loss": 0.1492, + "step": 39690 + }, + { + "epoch": 0.6192674861171773, + "grad_norm": 4.924708843231201, + "learning_rate": 0.000158716540837337, + "loss": 0.2234, + "step": 39700 + }, + { + "epoch": 0.619423472889499, + "grad_norm": 2.3538808822631836, + "learning_rate": 0.0001587061417191822, + "loss": 0.3405, + "step": 39710 + }, + { + "epoch": 0.6195794596618207, + "grad_norm": 0.24888922274112701, + "learning_rate": 0.00015869574260102745, + "loss": 0.1021, + "step": 39720 + }, + { + "epoch": 0.6197354464341424, + "grad_norm": 1.2719457149505615, + "learning_rate": 0.00015868534348287264, + "loss": 0.2832, + "step": 39730 + }, + { + "epoch": 0.6198914332064641, + "grad_norm": 2.80794358253479, + "learning_rate": 0.0001586749443647179, + "loss": 0.2295, + "step": 39740 + }, + { + "epoch": 0.6200474199787858, + "grad_norm": 0.09828778356313705, + "learning_rate": 0.00015866454524656308, + "loss": 0.3574, + "step": 39750 + }, + { + "epoch": 0.6202034067511075, + "grad_norm": 1.0614128112792969, + "learning_rate": 0.00015865414612840833, + "loss": 0.1886, + "step": 39760 + }, + { + "epoch": 0.6203593935234292, + "grad_norm": 0.6939805150032043, + "learning_rate": 0.00015864374701025352, + "loss": 0.3412, + "step": 39770 + }, + { + "epoch": 0.6205153802957509, + "grad_norm": 0.06902116537094116, + "learning_rate": 0.00015863334789209877, + "loss": 0.1647, + "step": 39780 + }, + { + "epoch": 0.6206713670680726, + "grad_norm": 2.05729079246521, + "learning_rate": 0.00015862294877394396, + "loss": 0.168, + "step": 39790 + }, + { + "epoch": 0.6208273538403943, + "grad_norm": 4.69445276260376, + "learning_rate": 0.0001586125496557892, + "loss": 0.1701, + "step": 39800 + }, + { + "epoch": 0.6209833406127161, + "grad_norm": 0.6905967593193054, + "learning_rate": 0.0001586021505376344, + "loss": 0.2823, + "step": 39810 + }, + { + "epoch": 0.6211393273850377, + "grad_norm": 3.0361857414245605, + "learning_rate": 0.00015859175141947964, + "loss": 0.1839, + "step": 39820 + }, + { + "epoch": 0.6212953141573595, + "grad_norm": 0.08303021639585495, + "learning_rate": 0.00015858135230132484, + "loss": 0.2183, + "step": 39830 + }, + { + "epoch": 0.6214513009296811, + "grad_norm": 3.230077028274536, + "learning_rate": 0.00015857095318317008, + "loss": 0.1727, + "step": 39840 + }, + { + "epoch": 0.6216072877020029, + "grad_norm": 1.2884387969970703, + "learning_rate": 0.00015856055406501528, + "loss": 0.1081, + "step": 39850 + }, + { + "epoch": 0.6217632744743246, + "grad_norm": 2.912475109100342, + "learning_rate": 0.00015855015494686052, + "loss": 0.2695, + "step": 39860 + }, + { + "epoch": 0.6219192612466463, + "grad_norm": 0.14122223854064941, + "learning_rate": 0.00015853975582870572, + "loss": 0.3006, + "step": 39870 + }, + { + "epoch": 0.622075248018968, + "grad_norm": 0.8160560131072998, + "learning_rate": 0.00015852935671055096, + "loss": 0.1204, + "step": 39880 + }, + { + "epoch": 0.6222312347912897, + "grad_norm": 1.6930943727493286, + "learning_rate": 0.00015851895759239615, + "loss": 0.1809, + "step": 39890 + }, + { + "epoch": 0.6223872215636114, + "grad_norm": 0.008987984620034695, + "learning_rate": 0.0001585085584742414, + "loss": 0.125, + "step": 39900 + }, + { + "epoch": 0.6225432083359331, + "grad_norm": 2.6068177223205566, + "learning_rate": 0.0001584981593560866, + "loss": 0.3131, + "step": 39910 + }, + { + "epoch": 0.6226991951082548, + "grad_norm": 0.5021041035652161, + "learning_rate": 0.00015848776023793184, + "loss": 0.2023, + "step": 39920 + }, + { + "epoch": 0.6228551818805765, + "grad_norm": 1.8558531999588013, + "learning_rate": 0.00015847736111977703, + "loss": 0.3135, + "step": 39930 + }, + { + "epoch": 0.6230111686528982, + "grad_norm": 0.019155239686369896, + "learning_rate": 0.00015846696200162228, + "loss": 0.2282, + "step": 39940 + }, + { + "epoch": 0.6231671554252199, + "grad_norm": 0.47993865609169006, + "learning_rate": 0.00015845656288346747, + "loss": 0.1143, + "step": 39950 + }, + { + "epoch": 0.6233231421975417, + "grad_norm": 2.553943157196045, + "learning_rate": 0.00015844616376531272, + "loss": 0.2061, + "step": 39960 + }, + { + "epoch": 0.6234791289698633, + "grad_norm": 0.22600245475769043, + "learning_rate": 0.0001584357646471579, + "loss": 0.1808, + "step": 39970 + }, + { + "epoch": 0.6236351157421851, + "grad_norm": 0.7153030037879944, + "learning_rate": 0.00015842536552900316, + "loss": 0.2221, + "step": 39980 + }, + { + "epoch": 0.6237911025145068, + "grad_norm": 0.12053301930427551, + "learning_rate": 0.00015841496641084835, + "loss": 0.2349, + "step": 39990 + }, + { + "epoch": 0.6239470892868285, + "grad_norm": 0.44607964158058167, + "learning_rate": 0.0001584045672926936, + "loss": 0.234, + "step": 40000 + }, + { + "epoch": 0.6241030760591502, + "grad_norm": 4.0730366706848145, + "learning_rate": 0.0001583941681745388, + "loss": 0.3213, + "step": 40010 + }, + { + "epoch": 0.6242590628314719, + "grad_norm": 2.8753201961517334, + "learning_rate": 0.00015838376905638404, + "loss": 0.1745, + "step": 40020 + }, + { + "epoch": 0.6244150496037936, + "grad_norm": 1.7546380758285522, + "learning_rate": 0.00015837336993822926, + "loss": 0.2564, + "step": 40030 + }, + { + "epoch": 0.6245710363761153, + "grad_norm": 2.7480337619781494, + "learning_rate": 0.00015836297082007448, + "loss": 0.2722, + "step": 40040 + }, + { + "epoch": 0.624727023148437, + "grad_norm": 1.9213544130325317, + "learning_rate": 0.0001583525717019197, + "loss": 0.3495, + "step": 40050 + }, + { + "epoch": 0.6248830099207587, + "grad_norm": 1.5569000244140625, + "learning_rate": 0.00015834217258376492, + "loss": 0.1744, + "step": 40060 + }, + { + "epoch": 0.6250389966930804, + "grad_norm": 0.4767601788043976, + "learning_rate": 0.00015833177346561014, + "loss": 0.1739, + "step": 40070 + }, + { + "epoch": 0.6251949834654021, + "grad_norm": 3.514345407485962, + "learning_rate": 0.00015832137434745536, + "loss": 0.3174, + "step": 40080 + }, + { + "epoch": 0.6253509702377238, + "grad_norm": 0.7669044137001038, + "learning_rate": 0.00015831097522930057, + "loss": 0.1893, + "step": 40090 + }, + { + "epoch": 0.6255069570100456, + "grad_norm": 0.3902212083339691, + "learning_rate": 0.0001583005761111458, + "loss": 0.2533, + "step": 40100 + }, + { + "epoch": 0.6256629437823673, + "grad_norm": 1.3150569200515747, + "learning_rate": 0.00015829017699299101, + "loss": 0.2491, + "step": 40110 + }, + { + "epoch": 0.625818930554689, + "grad_norm": 0.9833875894546509, + "learning_rate": 0.00015827977787483623, + "loss": 0.2889, + "step": 40120 + }, + { + "epoch": 0.6259749173270107, + "grad_norm": 1.0952039957046509, + "learning_rate": 0.00015826937875668145, + "loss": 0.2213, + "step": 40130 + }, + { + "epoch": 0.6261309040993324, + "grad_norm": 0.32410308718681335, + "learning_rate": 0.00015825897963852665, + "loss": 0.1123, + "step": 40140 + }, + { + "epoch": 0.6262868908716541, + "grad_norm": 0.05873742699623108, + "learning_rate": 0.0001582485805203719, + "loss": 0.1944, + "step": 40150 + }, + { + "epoch": 0.6264428776439758, + "grad_norm": 0.936841607093811, + "learning_rate": 0.00015823818140221709, + "loss": 0.2605, + "step": 40160 + }, + { + "epoch": 0.6265988644162975, + "grad_norm": 3.288151264190674, + "learning_rate": 0.00015822778228406233, + "loss": 0.201, + "step": 40170 + }, + { + "epoch": 0.6267548511886192, + "grad_norm": 0.45985645055770874, + "learning_rate": 0.00015821738316590752, + "loss": 0.094, + "step": 40180 + }, + { + "epoch": 0.6269108379609409, + "grad_norm": 3.8212318420410156, + "learning_rate": 0.00015820698404775277, + "loss": 0.2554, + "step": 40190 + }, + { + "epoch": 0.6270668247332626, + "grad_norm": 1.8019777536392212, + "learning_rate": 0.00015819658492959796, + "loss": 0.5479, + "step": 40200 + }, + { + "epoch": 0.6272228115055843, + "grad_norm": 0.49328944087028503, + "learning_rate": 0.0001581861858114432, + "loss": 0.2608, + "step": 40210 + }, + { + "epoch": 0.627378798277906, + "grad_norm": 0.30712059140205383, + "learning_rate": 0.0001581757866932884, + "loss": 0.235, + "step": 40220 + }, + { + "epoch": 0.6275347850502278, + "grad_norm": 3.460827112197876, + "learning_rate": 0.00015816538757513365, + "loss": 0.3527, + "step": 40230 + }, + { + "epoch": 0.6276907718225494, + "grad_norm": 1.6247568130493164, + "learning_rate": 0.00015815498845697884, + "loss": 0.3031, + "step": 40240 + }, + { + "epoch": 0.6278467585948712, + "grad_norm": 0.5068777203559875, + "learning_rate": 0.0001581445893388241, + "loss": 0.3529, + "step": 40250 + }, + { + "epoch": 0.6280027453671929, + "grad_norm": 0.2704099416732788, + "learning_rate": 0.00015813419022066928, + "loss": 0.1043, + "step": 40260 + }, + { + "epoch": 0.6281587321395146, + "grad_norm": 0.07768986374139786, + "learning_rate": 0.00015812379110251453, + "loss": 0.257, + "step": 40270 + }, + { + "epoch": 0.6283147189118363, + "grad_norm": 2.6940250396728516, + "learning_rate": 0.00015811339198435972, + "loss": 0.1694, + "step": 40280 + }, + { + "epoch": 0.628470705684158, + "grad_norm": 2.3470752239227295, + "learning_rate": 0.00015810299286620497, + "loss": 0.2196, + "step": 40290 + }, + { + "epoch": 0.6286266924564797, + "grad_norm": 1.6333197355270386, + "learning_rate": 0.00015809259374805016, + "loss": 0.2452, + "step": 40300 + }, + { + "epoch": 0.6287826792288014, + "grad_norm": 0.28753232955932617, + "learning_rate": 0.0001580821946298954, + "loss": 0.0601, + "step": 40310 + }, + { + "epoch": 0.6289386660011231, + "grad_norm": 0.8923532366752625, + "learning_rate": 0.0001580717955117406, + "loss": 0.455, + "step": 40320 + }, + { + "epoch": 0.6290946527734448, + "grad_norm": 2.404088020324707, + "learning_rate": 0.00015806139639358585, + "loss": 0.466, + "step": 40330 + }, + { + "epoch": 0.6292506395457665, + "grad_norm": 1.0499215126037598, + "learning_rate": 0.00015805099727543104, + "loss": 0.4372, + "step": 40340 + }, + { + "epoch": 0.6294066263180882, + "grad_norm": 0.21907389163970947, + "learning_rate": 0.00015804059815727629, + "loss": 0.1628, + "step": 40350 + }, + { + "epoch": 0.62956261309041, + "grad_norm": 1.1540305614471436, + "learning_rate": 0.00015803019903912148, + "loss": 0.1888, + "step": 40360 + }, + { + "epoch": 0.6297185998627316, + "grad_norm": 0.44641104340553284, + "learning_rate": 0.00015801979992096672, + "loss": 0.2372, + "step": 40370 + }, + { + "epoch": 0.6298745866350534, + "grad_norm": 1.5536185503005981, + "learning_rate": 0.00015800940080281192, + "loss": 0.0855, + "step": 40380 + }, + { + "epoch": 0.630030573407375, + "grad_norm": 5.066720485687256, + "learning_rate": 0.00015799900168465716, + "loss": 0.3324, + "step": 40390 + }, + { + "epoch": 0.6301865601796968, + "grad_norm": 1.0257264375686646, + "learning_rate": 0.00015798860256650236, + "loss": 0.1872, + "step": 40400 + }, + { + "epoch": 0.6303425469520185, + "grad_norm": 1.108614206314087, + "learning_rate": 0.0001579782034483476, + "loss": 0.3244, + "step": 40410 + }, + { + "epoch": 0.6304985337243402, + "grad_norm": 0.9430763125419617, + "learning_rate": 0.0001579678043301928, + "loss": 0.2388, + "step": 40420 + }, + { + "epoch": 0.6306545204966619, + "grad_norm": 0.3578563332557678, + "learning_rate": 0.00015795740521203804, + "loss": 0.3072, + "step": 40430 + }, + { + "epoch": 0.6308105072689836, + "grad_norm": 0.6251798272132874, + "learning_rate": 0.00015794700609388323, + "loss": 0.3246, + "step": 40440 + }, + { + "epoch": 0.6309664940413053, + "grad_norm": 1.5441378355026245, + "learning_rate": 0.00015793660697572848, + "loss": 0.1822, + "step": 40450 + }, + { + "epoch": 0.631122480813627, + "grad_norm": 1.3051613569259644, + "learning_rate": 0.00015792620785757367, + "loss": 0.3554, + "step": 40460 + }, + { + "epoch": 0.6312784675859487, + "grad_norm": 4.274336338043213, + "learning_rate": 0.00015791580873941892, + "loss": 0.2423, + "step": 40470 + }, + { + "epoch": 0.6314344543582704, + "grad_norm": 1.4855650663375854, + "learning_rate": 0.0001579054096212641, + "loss": 0.2158, + "step": 40480 + }, + { + "epoch": 0.6315904411305922, + "grad_norm": 0.26689472794532776, + "learning_rate": 0.00015789501050310936, + "loss": 0.3227, + "step": 40490 + }, + { + "epoch": 0.6317464279029138, + "grad_norm": 2.0525012016296387, + "learning_rate": 0.00015788461138495455, + "loss": 0.2475, + "step": 40500 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.466279456768e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-40500/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..16327c268b80e000346b82cfd40d4b2cd6ede828 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cba145ff393431d11a073c793084c5c816831e219237f2833a67b451fb8b4908 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..3217531248c1a3e1f9b0971ffb8dee53e558abbb --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:166a971f0e313ac5c6a77c652e3bedb3b5c35a03a9fb45b42b747ff39eeb50b2 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..96f63b29f9cfc3ce5854b729d358bf90cd0a0df8 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..216b9f866519f41b365f70ddabf5415342284672 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..755fba05842649b27c67f45fd06c943946a83392 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/trainer_state.json @@ -0,0 +1,28734 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.6395457665189992, + "eval_steps": 500, + "global_step": 41000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + }, + { + "epoch": 0.5071129968178698, + "grad_norm": 1.0075199604034424, + "learning_rate": 0.00016619350679062414, + "loss": 0.2569, + "step": 32510 + }, + { + "epoch": 0.5072689835901916, + "grad_norm": 1.3489066362380981, + "learning_rate": 0.0001661831076724694, + "loss": 0.2565, + "step": 32520 + }, + { + "epoch": 0.5074249703625132, + "grad_norm": 0.2534058094024658, + "learning_rate": 0.00016617270855431458, + "loss": 0.2292, + "step": 32530 + }, + { + "epoch": 0.507580957134835, + "grad_norm": 0.31723281741142273, + "learning_rate": 0.00016616230943615983, + "loss": 0.1467, + "step": 32540 + }, + { + "epoch": 0.5077369439071566, + "grad_norm": 0.29659849405288696, + "learning_rate": 0.00016615191031800502, + "loss": 0.2768, + "step": 32550 + }, + { + "epoch": 0.5078929306794784, + "grad_norm": 1.5246435403823853, + "learning_rate": 0.00016614151119985027, + "loss": 0.216, + "step": 32560 + }, + { + "epoch": 0.5080489174518001, + "grad_norm": 0.6026854515075684, + "learning_rate": 0.00016613111208169546, + "loss": 0.2277, + "step": 32570 + }, + { + "epoch": 0.5082049042241218, + "grad_norm": 0.41994890570640564, + "learning_rate": 0.0001661207129635407, + "loss": 0.2433, + "step": 32580 + }, + { + "epoch": 0.5083608909964435, + "grad_norm": 2.084599733352661, + "learning_rate": 0.00016611031384538593, + "loss": 0.4081, + "step": 32590 + }, + { + "epoch": 0.5085168777687652, + "grad_norm": 0.2945229709148407, + "learning_rate": 0.00016609991472723115, + "loss": 0.1329, + "step": 32600 + }, + { + "epoch": 0.508672864541087, + "grad_norm": 1.4981423616409302, + "learning_rate": 0.00016608951560907636, + "loss": 0.2855, + "step": 32610 + }, + { + "epoch": 0.5088288513134086, + "grad_norm": 0.961737871170044, + "learning_rate": 0.00016607911649092158, + "loss": 0.1654, + "step": 32620 + }, + { + "epoch": 0.5089848380857304, + "grad_norm": 1.49734628200531, + "learning_rate": 0.0001660687173727668, + "loss": 0.2081, + "step": 32630 + }, + { + "epoch": 0.509140824858052, + "grad_norm": 2.538135051727295, + "learning_rate": 0.00016605831825461202, + "loss": 0.3023, + "step": 32640 + }, + { + "epoch": 0.5092968116303738, + "grad_norm": 0.2981272041797638, + "learning_rate": 0.00016604791913645724, + "loss": 0.214, + "step": 32650 + }, + { + "epoch": 0.5094527984026954, + "grad_norm": 1.8863650560379028, + "learning_rate": 0.00016603752001830246, + "loss": 0.3363, + "step": 32660 + }, + { + "epoch": 0.5096087851750172, + "grad_norm": 0.022720983251929283, + "learning_rate": 0.00016602712090014768, + "loss": 0.2463, + "step": 32670 + }, + { + "epoch": 0.5097647719473388, + "grad_norm": 1.154977798461914, + "learning_rate": 0.0001660167217819929, + "loss": 0.4484, + "step": 32680 + }, + { + "epoch": 0.5099207587196606, + "grad_norm": 0.9470499157905579, + "learning_rate": 0.00016600632266383812, + "loss": 0.1196, + "step": 32690 + }, + { + "epoch": 0.5100767454919822, + "grad_norm": 0.19955401122570038, + "learning_rate": 0.00016599592354568334, + "loss": 0.1762, + "step": 32700 + }, + { + "epoch": 0.510232732264304, + "grad_norm": 0.4098282754421234, + "learning_rate": 0.00016598552442752856, + "loss": 0.2739, + "step": 32710 + }, + { + "epoch": 0.5103887190366257, + "grad_norm": 0.21697258949279785, + "learning_rate": 0.00016597512530937378, + "loss": 0.2555, + "step": 32720 + }, + { + "epoch": 0.5105447058089474, + "grad_norm": 4.894479751586914, + "learning_rate": 0.000165964726191219, + "loss": 0.3137, + "step": 32730 + }, + { + "epoch": 0.5107006925812692, + "grad_norm": 0.8396711349487305, + "learning_rate": 0.00016595432707306422, + "loss": 0.3105, + "step": 32740 + }, + { + "epoch": 0.5108566793535908, + "grad_norm": 1.5494674444198608, + "learning_rate": 0.00016594392795490944, + "loss": 0.2271, + "step": 32750 + }, + { + "epoch": 0.5110126661259126, + "grad_norm": 1.3409888744354248, + "learning_rate": 0.00016593352883675466, + "loss": 0.1796, + "step": 32760 + }, + { + "epoch": 0.5111686528982342, + "grad_norm": 1.7590911388397217, + "learning_rate": 0.00016592312971859988, + "loss": 0.2185, + "step": 32770 + }, + { + "epoch": 0.511324639670556, + "grad_norm": 0.7350002527236938, + "learning_rate": 0.0001659127306004451, + "loss": 0.1911, + "step": 32780 + }, + { + "epoch": 0.5114806264428776, + "grad_norm": 0.7176411747932434, + "learning_rate": 0.00016590233148229032, + "loss": 0.6918, + "step": 32790 + }, + { + "epoch": 0.5116366132151994, + "grad_norm": 2.5005321502685547, + "learning_rate": 0.00016589193236413554, + "loss": 0.2329, + "step": 32800 + }, + { + "epoch": 0.511792599987521, + "grad_norm": 1.3916250467300415, + "learning_rate": 0.00016588153324598076, + "loss": 0.2102, + "step": 32810 + }, + { + "epoch": 0.5119485867598428, + "grad_norm": 0.721671998500824, + "learning_rate": 0.00016587113412782598, + "loss": 0.2965, + "step": 32820 + }, + { + "epoch": 0.5121045735321644, + "grad_norm": 0.6120114326477051, + "learning_rate": 0.0001658607350096712, + "loss": 0.2661, + "step": 32830 + }, + { + "epoch": 0.5122605603044862, + "grad_norm": 0.29123347997665405, + "learning_rate": 0.00016585033589151642, + "loss": 0.2455, + "step": 32840 + }, + { + "epoch": 0.5124165470768078, + "grad_norm": 2.7300868034362793, + "learning_rate": 0.00016583993677336164, + "loss": 0.2049, + "step": 32850 + }, + { + "epoch": 0.5125725338491296, + "grad_norm": 1.801928997039795, + "learning_rate": 0.00016582953765520686, + "loss": 0.1725, + "step": 32860 + }, + { + "epoch": 0.5127285206214514, + "grad_norm": 0.6406471729278564, + "learning_rate": 0.00016581913853705208, + "loss": 0.164, + "step": 32870 + }, + { + "epoch": 0.512884507393773, + "grad_norm": 0.9627325534820557, + "learning_rate": 0.0001658087394188973, + "loss": 0.1728, + "step": 32880 + }, + { + "epoch": 0.5130404941660948, + "grad_norm": 1.5446962118148804, + "learning_rate": 0.00016579834030074251, + "loss": 0.2345, + "step": 32890 + }, + { + "epoch": 0.5131964809384164, + "grad_norm": 0.28092706203460693, + "learning_rate": 0.0001657879411825877, + "loss": 0.2124, + "step": 32900 + }, + { + "epoch": 0.5133524677107382, + "grad_norm": 1.7751421928405762, + "learning_rate": 0.00016577754206443295, + "loss": 0.2367, + "step": 32910 + }, + { + "epoch": 0.5135084544830598, + "grad_norm": 1.6964634656906128, + "learning_rate": 0.00016576714294627815, + "loss": 0.2361, + "step": 32920 + }, + { + "epoch": 0.5136644412553816, + "grad_norm": 0.06759823113679886, + "learning_rate": 0.0001657567438281234, + "loss": 0.2329, + "step": 32930 + }, + { + "epoch": 0.5138204280277032, + "grad_norm": 2.2519280910491943, + "learning_rate": 0.00016574634470996859, + "loss": 0.1079, + "step": 32940 + }, + { + "epoch": 0.513976414800025, + "grad_norm": 3.167576551437378, + "learning_rate": 0.00016573594559181383, + "loss": 0.1764, + "step": 32950 + }, + { + "epoch": 0.5141324015723466, + "grad_norm": 1.6734397411346436, + "learning_rate": 0.00016572554647365903, + "loss": 0.2878, + "step": 32960 + }, + { + "epoch": 0.5142883883446684, + "grad_norm": 0.35627374053001404, + "learning_rate": 0.00016571514735550427, + "loss": 0.2337, + "step": 32970 + }, + { + "epoch": 0.51444437511699, + "grad_norm": 1.6321039199829102, + "learning_rate": 0.00016570474823734946, + "loss": 0.2551, + "step": 32980 + }, + { + "epoch": 0.5146003618893118, + "grad_norm": 1.352982997894287, + "learning_rate": 0.0001656943491191947, + "loss": 0.4467, + "step": 32990 + }, + { + "epoch": 0.5147563486616334, + "grad_norm": 0.3558129072189331, + "learning_rate": 0.0001656839500010399, + "loss": 0.1913, + "step": 33000 + }, + { + "epoch": 0.5149123354339552, + "grad_norm": 0.041972286999225616, + "learning_rate": 0.00016567355088288515, + "loss": 0.1704, + "step": 33010 + }, + { + "epoch": 0.515068322206277, + "grad_norm": 1.4168939590454102, + "learning_rate": 0.00016566315176473034, + "loss": 0.4866, + "step": 33020 + }, + { + "epoch": 0.5152243089785986, + "grad_norm": 3.0042929649353027, + "learning_rate": 0.0001656527526465756, + "loss": 0.2473, + "step": 33030 + }, + { + "epoch": 0.5153802957509204, + "grad_norm": 1.027274489402771, + "learning_rate": 0.00016564235352842078, + "loss": 0.436, + "step": 33040 + }, + { + "epoch": 0.515536282523242, + "grad_norm": 1.2330195903778076, + "learning_rate": 0.00016563195441026603, + "loss": 0.4615, + "step": 33050 + }, + { + "epoch": 0.5156922692955638, + "grad_norm": 2.7145066261291504, + "learning_rate": 0.00016562155529211122, + "loss": 0.278, + "step": 33060 + }, + { + "epoch": 0.5158482560678854, + "grad_norm": 1.1273893117904663, + "learning_rate": 0.00016561115617395647, + "loss": 0.3767, + "step": 33070 + }, + { + "epoch": 0.5160042428402072, + "grad_norm": 2.466265916824341, + "learning_rate": 0.00016560075705580166, + "loss": 0.1879, + "step": 33080 + }, + { + "epoch": 0.5161602296125288, + "grad_norm": 0.3759515583515167, + "learning_rate": 0.0001655903579376469, + "loss": 0.4004, + "step": 33090 + }, + { + "epoch": 0.5163162163848506, + "grad_norm": 1.8821336030960083, + "learning_rate": 0.0001655799588194921, + "loss": 0.1622, + "step": 33100 + }, + { + "epoch": 0.5164722031571722, + "grad_norm": 0.8162096738815308, + "learning_rate": 0.00016556955970133735, + "loss": 0.1341, + "step": 33110 + }, + { + "epoch": 0.516628189929494, + "grad_norm": 0.008498801849782467, + "learning_rate": 0.00016555916058318254, + "loss": 0.2546, + "step": 33120 + }, + { + "epoch": 0.5167841767018156, + "grad_norm": 0.17536675930023193, + "learning_rate": 0.00016554876146502779, + "loss": 0.2934, + "step": 33130 + }, + { + "epoch": 0.5169401634741374, + "grad_norm": 3.539898633956909, + "learning_rate": 0.00016553836234687298, + "loss": 0.4959, + "step": 33140 + }, + { + "epoch": 0.517096150246459, + "grad_norm": 2.190887212753296, + "learning_rate": 0.00016552796322871823, + "loss": 0.2914, + "step": 33150 + }, + { + "epoch": 0.5172521370187808, + "grad_norm": 1.6388943195343018, + "learning_rate": 0.00016551756411056342, + "loss": 0.1317, + "step": 33160 + }, + { + "epoch": 0.5174081237911026, + "grad_norm": 0.33063387870788574, + "learning_rate": 0.00016550716499240866, + "loss": 0.4258, + "step": 33170 + }, + { + "epoch": 0.5175641105634242, + "grad_norm": 0.5168288946151733, + "learning_rate": 0.00016549676587425386, + "loss": 0.2548, + "step": 33180 + }, + { + "epoch": 0.517720097335746, + "grad_norm": 3.9570281505584717, + "learning_rate": 0.0001654863667560991, + "loss": 0.3179, + "step": 33190 + }, + { + "epoch": 0.5178760841080676, + "grad_norm": 1.127255916595459, + "learning_rate": 0.0001654759676379443, + "loss": 0.233, + "step": 33200 + }, + { + "epoch": 0.5180320708803894, + "grad_norm": 0.36299410462379456, + "learning_rate": 0.00016546556851978954, + "loss": 0.4614, + "step": 33210 + }, + { + "epoch": 0.518188057652711, + "grad_norm": 2.914668321609497, + "learning_rate": 0.00016545516940163474, + "loss": 0.1603, + "step": 33220 + }, + { + "epoch": 0.5183440444250328, + "grad_norm": 6.991094589233398, + "learning_rate": 0.00016544477028347998, + "loss": 0.2259, + "step": 33230 + }, + { + "epoch": 0.5185000311973544, + "grad_norm": 2.0594241619110107, + "learning_rate": 0.00016543437116532517, + "loss": 0.1885, + "step": 33240 + }, + { + "epoch": 0.5186560179696762, + "grad_norm": 0.5594693422317505, + "learning_rate": 0.00016542397204717042, + "loss": 0.4333, + "step": 33250 + }, + { + "epoch": 0.5188120047419978, + "grad_norm": 4.314854621887207, + "learning_rate": 0.00016541357292901561, + "loss": 0.2103, + "step": 33260 + }, + { + "epoch": 0.5189679915143196, + "grad_norm": 0.5609000325202942, + "learning_rate": 0.00016540317381086086, + "loss": 0.2498, + "step": 33270 + }, + { + "epoch": 0.5191239782866413, + "grad_norm": 3.680088758468628, + "learning_rate": 0.00016539277469270605, + "loss": 0.3465, + "step": 33280 + }, + { + "epoch": 0.519279965058963, + "grad_norm": 1.4993672370910645, + "learning_rate": 0.0001653823755745513, + "loss": 0.3272, + "step": 33290 + }, + { + "epoch": 0.5194359518312847, + "grad_norm": 0.18005084991455078, + "learning_rate": 0.0001653719764563965, + "loss": 0.1706, + "step": 33300 + }, + { + "epoch": 0.5195919386036064, + "grad_norm": 0.5105524659156799, + "learning_rate": 0.00016536157733824174, + "loss": 0.0937, + "step": 33310 + }, + { + "epoch": 0.5197479253759281, + "grad_norm": 1.2806589603424072, + "learning_rate": 0.00016535117822008693, + "loss": 0.1578, + "step": 33320 + }, + { + "epoch": 0.5199039121482498, + "grad_norm": 0.22015990316867828, + "learning_rate": 0.00016534077910193218, + "loss": 0.2081, + "step": 33330 + }, + { + "epoch": 0.5200598989205716, + "grad_norm": 0.34638991951942444, + "learning_rate": 0.00016533037998377737, + "loss": 0.2389, + "step": 33340 + }, + { + "epoch": 0.5202158856928932, + "grad_norm": 0.8709302544593811, + "learning_rate": 0.0001653199808656226, + "loss": 0.2112, + "step": 33350 + }, + { + "epoch": 0.520371872465215, + "grad_norm": 1.2309521436691284, + "learning_rate": 0.0001653095817474678, + "loss": 0.3448, + "step": 33360 + }, + { + "epoch": 0.5205278592375366, + "grad_norm": 1.3155115842819214, + "learning_rate": 0.00016529918262931303, + "loss": 0.1572, + "step": 33370 + }, + { + "epoch": 0.5206838460098584, + "grad_norm": 0.358751118183136, + "learning_rate": 0.00016528878351115825, + "loss": 0.3011, + "step": 33380 + }, + { + "epoch": 0.52083983278218, + "grad_norm": 3.800628662109375, + "learning_rate": 0.00016527838439300347, + "loss": 0.458, + "step": 33390 + }, + { + "epoch": 0.5209958195545018, + "grad_norm": 1.1253124475479126, + "learning_rate": 0.0001652679852748487, + "loss": 0.1238, + "step": 33400 + }, + { + "epoch": 0.5211518063268235, + "grad_norm": 0.46979108452796936, + "learning_rate": 0.0001652575861566939, + "loss": 0.0745, + "step": 33410 + }, + { + "epoch": 0.5213077930991452, + "grad_norm": 7.921492576599121, + "learning_rate": 0.00016524718703853913, + "loss": 0.4004, + "step": 33420 + }, + { + "epoch": 0.5214637798714669, + "grad_norm": 1.643127679824829, + "learning_rate": 0.00016523678792038435, + "loss": 0.2858, + "step": 33430 + }, + { + "epoch": 0.5216197666437886, + "grad_norm": 7.096631050109863, + "learning_rate": 0.0001652263888022296, + "loss": 0.4735, + "step": 33440 + }, + { + "epoch": 0.5217757534161103, + "grad_norm": 0.7280219197273254, + "learning_rate": 0.0001652159896840748, + "loss": 0.1841, + "step": 33450 + }, + { + "epoch": 0.521931740188432, + "grad_norm": 0.9415522813796997, + "learning_rate": 0.00016520559056592003, + "loss": 0.2169, + "step": 33460 + }, + { + "epoch": 0.5220877269607537, + "grad_norm": 2.264862060546875, + "learning_rate": 0.00016519519144776523, + "loss": 0.1902, + "step": 33470 + }, + { + "epoch": 0.5222437137330754, + "grad_norm": 2.4566822052001953, + "learning_rate": 0.00016518479232961047, + "loss": 0.3329, + "step": 33480 + }, + { + "epoch": 0.5223997005053972, + "grad_norm": 0.7330378890037537, + "learning_rate": 0.00016517439321145567, + "loss": 0.2341, + "step": 33490 + }, + { + "epoch": 0.5225556872777188, + "grad_norm": 1.6409542560577393, + "learning_rate": 0.0001651639940933009, + "loss": 0.3077, + "step": 33500 + }, + { + "epoch": 0.5227116740500406, + "grad_norm": 1.0647432804107666, + "learning_rate": 0.0001651535949751461, + "loss": 0.1935, + "step": 33510 + }, + { + "epoch": 0.5228676608223622, + "grad_norm": 0.507420539855957, + "learning_rate": 0.00016514319585699135, + "loss": 0.1314, + "step": 33520 + }, + { + "epoch": 0.523023647594684, + "grad_norm": 1.0064164400100708, + "learning_rate": 0.00016513279673883654, + "loss": 0.344, + "step": 33530 + }, + { + "epoch": 0.5231796343670057, + "grad_norm": 1.1936209201812744, + "learning_rate": 0.0001651223976206818, + "loss": 0.1948, + "step": 33540 + }, + { + "epoch": 0.5233356211393274, + "grad_norm": 0.04992926865816116, + "learning_rate": 0.00016511199850252698, + "loss": 0.1363, + "step": 33550 + }, + { + "epoch": 0.5234916079116491, + "grad_norm": 1.8256275653839111, + "learning_rate": 0.00016510159938437223, + "loss": 0.443, + "step": 33560 + }, + { + "epoch": 0.5236475946839708, + "grad_norm": 4.914908409118652, + "learning_rate": 0.00016509120026621742, + "loss": 0.4654, + "step": 33570 + }, + { + "epoch": 0.5238035814562925, + "grad_norm": 1.1625486612319946, + "learning_rate": 0.00016508080114806267, + "loss": 0.1974, + "step": 33580 + }, + { + "epoch": 0.5239595682286142, + "grad_norm": 6.6192522048950195, + "learning_rate": 0.00016507040202990786, + "loss": 0.2131, + "step": 33590 + }, + { + "epoch": 0.5241155550009359, + "grad_norm": 0.7369065880775452, + "learning_rate": 0.0001650600029117531, + "loss": 0.2371, + "step": 33600 + }, + { + "epoch": 0.5242715417732576, + "grad_norm": 1.5238152742385864, + "learning_rate": 0.0001650496037935983, + "loss": 0.1557, + "step": 33610 + }, + { + "epoch": 0.5244275285455793, + "grad_norm": 1.0418007373809814, + "learning_rate": 0.00016503920467544355, + "loss": 0.1878, + "step": 33620 + }, + { + "epoch": 0.524583515317901, + "grad_norm": 0.790117084980011, + "learning_rate": 0.00016502880555728874, + "loss": 0.2195, + "step": 33630 + }, + { + "epoch": 0.5247395020902228, + "grad_norm": 1.6712257862091064, + "learning_rate": 0.000165018406439134, + "loss": 0.1602, + "step": 33640 + }, + { + "epoch": 0.5248954888625444, + "grad_norm": 0.19236230850219727, + "learning_rate": 0.00016500800732097918, + "loss": 0.2526, + "step": 33650 + }, + { + "epoch": 0.5250514756348662, + "grad_norm": 1.3519701957702637, + "learning_rate": 0.00016499760820282443, + "loss": 0.4686, + "step": 33660 + }, + { + "epoch": 0.5252074624071879, + "grad_norm": 1.694342851638794, + "learning_rate": 0.00016498720908466962, + "loss": 0.1859, + "step": 33670 + }, + { + "epoch": 0.5253634491795096, + "grad_norm": 5.225239276885986, + "learning_rate": 0.00016497680996651487, + "loss": 0.2944, + "step": 33680 + }, + { + "epoch": 0.5255194359518313, + "grad_norm": 2.0208842754364014, + "learning_rate": 0.00016496641084836006, + "loss": 0.2421, + "step": 33690 + }, + { + "epoch": 0.525675422724153, + "grad_norm": 0.7954996228218079, + "learning_rate": 0.0001649560117302053, + "loss": 0.2853, + "step": 33700 + }, + { + "epoch": 0.5258314094964747, + "grad_norm": 2.296086072921753, + "learning_rate": 0.0001649456126120505, + "loss": 0.1333, + "step": 33710 + }, + { + "epoch": 0.5259873962687964, + "grad_norm": 1.1779128313064575, + "learning_rate": 0.00016493521349389574, + "loss": 0.1066, + "step": 33720 + }, + { + "epoch": 0.5261433830411181, + "grad_norm": 0.1756065934896469, + "learning_rate": 0.00016492481437574094, + "loss": 0.1352, + "step": 33730 + }, + { + "epoch": 0.5262993698134398, + "grad_norm": 0.13100725412368774, + "learning_rate": 0.00016491441525758618, + "loss": 0.2399, + "step": 33740 + }, + { + "epoch": 0.5264553565857615, + "grad_norm": 5.532008171081543, + "learning_rate": 0.00016490401613943138, + "loss": 0.2896, + "step": 33750 + }, + { + "epoch": 0.5266113433580832, + "grad_norm": 1.319886565208435, + "learning_rate": 0.00016489361702127662, + "loss": 0.3275, + "step": 33760 + }, + { + "epoch": 0.5267673301304049, + "grad_norm": 1.5550974607467651, + "learning_rate": 0.00016488321790312182, + "loss": 0.2677, + "step": 33770 + }, + { + "epoch": 0.5269233169027266, + "grad_norm": 1.8936737775802612, + "learning_rate": 0.00016487281878496706, + "loss": 0.1955, + "step": 33780 + }, + { + "epoch": 0.5270793036750484, + "grad_norm": 0.3653401732444763, + "learning_rate": 0.00016486241966681226, + "loss": 0.0723, + "step": 33790 + }, + { + "epoch": 0.52723529044737, + "grad_norm": 2.861341714859009, + "learning_rate": 0.00016485202054865747, + "loss": 0.2412, + "step": 33800 + }, + { + "epoch": 0.5273912772196918, + "grad_norm": 1.5291428565979004, + "learning_rate": 0.0001648416214305027, + "loss": 0.0871, + "step": 33810 + }, + { + "epoch": 0.5275472639920135, + "grad_norm": 1.0372581481933594, + "learning_rate": 0.00016483122231234791, + "loss": 0.4705, + "step": 33820 + }, + { + "epoch": 0.5277032507643352, + "grad_norm": 1.1943141222000122, + "learning_rate": 0.00016482082319419313, + "loss": 0.2848, + "step": 33830 + }, + { + "epoch": 0.5278592375366569, + "grad_norm": 1.9008225202560425, + "learning_rate": 0.00016481042407603835, + "loss": 0.139, + "step": 33840 + }, + { + "epoch": 0.5280152243089786, + "grad_norm": 2.132089138031006, + "learning_rate": 0.00016480002495788357, + "loss": 0.2119, + "step": 33850 + }, + { + "epoch": 0.5281712110813003, + "grad_norm": 0.24524426460266113, + "learning_rate": 0.0001647896258397288, + "loss": 0.1643, + "step": 33860 + }, + { + "epoch": 0.528327197853622, + "grad_norm": 1.6469637155532837, + "learning_rate": 0.000164779226721574, + "loss": 0.363, + "step": 33870 + }, + { + "epoch": 0.5284831846259437, + "grad_norm": 0.8767328858375549, + "learning_rate": 0.00016476882760341923, + "loss": 0.2632, + "step": 33880 + }, + { + "epoch": 0.5286391713982654, + "grad_norm": 0.06347586214542389, + "learning_rate": 0.00016475842848526445, + "loss": 0.3204, + "step": 33890 + }, + { + "epoch": 0.5287951581705871, + "grad_norm": 0.09782540798187256, + "learning_rate": 0.00016474802936710967, + "loss": 0.2211, + "step": 33900 + }, + { + "epoch": 0.5289511449429088, + "grad_norm": 3.2998859882354736, + "learning_rate": 0.0001647376302489549, + "loss": 0.265, + "step": 33910 + }, + { + "epoch": 0.5291071317152305, + "grad_norm": 0.43594226241111755, + "learning_rate": 0.0001647272311308001, + "loss": 0.194, + "step": 33920 + }, + { + "epoch": 0.5292631184875523, + "grad_norm": 1.5166605710983276, + "learning_rate": 0.00016471683201264533, + "loss": 0.2675, + "step": 33930 + }, + { + "epoch": 0.529419105259874, + "grad_norm": 0.6056640148162842, + "learning_rate": 0.00016470643289449055, + "loss": 0.0576, + "step": 33940 + }, + { + "epoch": 0.5295750920321957, + "grad_norm": 0.25410348176956177, + "learning_rate": 0.00016469603377633577, + "loss": 0.1124, + "step": 33950 + }, + { + "epoch": 0.5297310788045174, + "grad_norm": 1.770642876625061, + "learning_rate": 0.000164685634658181, + "loss": 0.3295, + "step": 33960 + }, + { + "epoch": 0.5298870655768391, + "grad_norm": 0.0607205331325531, + "learning_rate": 0.0001646752355400262, + "loss": 0.2369, + "step": 33970 + }, + { + "epoch": 0.5300430523491608, + "grad_norm": 0.5557095408439636, + "learning_rate": 0.00016466483642187143, + "loss": 0.1681, + "step": 33980 + }, + { + "epoch": 0.5301990391214825, + "grad_norm": 0.5192957520484924, + "learning_rate": 0.00016465443730371665, + "loss": 0.2649, + "step": 33990 + }, + { + "epoch": 0.5303550258938042, + "grad_norm": 0.04804835096001625, + "learning_rate": 0.00016464403818556187, + "loss": 0.2639, + "step": 34000 + }, + { + "epoch": 0.5305110126661259, + "grad_norm": 0.02673129364848137, + "learning_rate": 0.0001646336390674071, + "loss": 0.169, + "step": 34010 + }, + { + "epoch": 0.5306669994384476, + "grad_norm": 1.0084244012832642, + "learning_rate": 0.0001646232399492523, + "loss": 0.353, + "step": 34020 + }, + { + "epoch": 0.5308229862107693, + "grad_norm": 2.2202091217041016, + "learning_rate": 0.00016461284083109753, + "loss": 0.3206, + "step": 34030 + }, + { + "epoch": 0.530978972983091, + "grad_norm": 0.5573744773864746, + "learning_rate": 0.00016460244171294275, + "loss": 0.231, + "step": 34040 + }, + { + "epoch": 0.5311349597554127, + "grad_norm": 0.39700084924697876, + "learning_rate": 0.00016459204259478797, + "loss": 0.1412, + "step": 34050 + }, + { + "epoch": 0.5312909465277345, + "grad_norm": 2.582963228225708, + "learning_rate": 0.00016458164347663319, + "loss": 0.3334, + "step": 34060 + }, + { + "epoch": 0.5314469333000561, + "grad_norm": 0.22781169414520264, + "learning_rate": 0.0001645712443584784, + "loss": 0.2767, + "step": 34070 + }, + { + "epoch": 0.5316029200723779, + "grad_norm": 1.2051042318344116, + "learning_rate": 0.00016456084524032362, + "loss": 0.2045, + "step": 34080 + }, + { + "epoch": 0.5317589068446996, + "grad_norm": 0.42760100960731506, + "learning_rate": 0.00016455044612216884, + "loss": 0.1151, + "step": 34090 + }, + { + "epoch": 0.5319148936170213, + "grad_norm": 0.14440476894378662, + "learning_rate": 0.00016454004700401406, + "loss": 0.2386, + "step": 34100 + }, + { + "epoch": 0.532070880389343, + "grad_norm": 2.2777981758117676, + "learning_rate": 0.00016452964788585928, + "loss": 0.2009, + "step": 34110 + }, + { + "epoch": 0.5322268671616647, + "grad_norm": 0.9206979274749756, + "learning_rate": 0.0001645192487677045, + "loss": 0.2745, + "step": 34120 + }, + { + "epoch": 0.5323828539339864, + "grad_norm": 1.6947574615478516, + "learning_rate": 0.00016450884964954972, + "loss": 0.2584, + "step": 34130 + }, + { + "epoch": 0.5325388407063081, + "grad_norm": 0.401444673538208, + "learning_rate": 0.00016449845053139494, + "loss": 0.2218, + "step": 34140 + }, + { + "epoch": 0.5326948274786298, + "grad_norm": 0.08261553198099136, + "learning_rate": 0.00016448805141324016, + "loss": 0.2775, + "step": 34150 + }, + { + "epoch": 0.5328508142509515, + "grad_norm": 0.1017974391579628, + "learning_rate": 0.00016447765229508538, + "loss": 0.2095, + "step": 34160 + }, + { + "epoch": 0.5330068010232732, + "grad_norm": 1.3759571313858032, + "learning_rate": 0.0001644672531769306, + "loss": 0.2643, + "step": 34170 + }, + { + "epoch": 0.5331627877955949, + "grad_norm": 1.2654389142990112, + "learning_rate": 0.00016445685405877582, + "loss": 0.2949, + "step": 34180 + }, + { + "epoch": 0.5333187745679167, + "grad_norm": 1.5481843948364258, + "learning_rate": 0.00016444645494062104, + "loss": 0.1264, + "step": 34190 + }, + { + "epoch": 0.5334747613402383, + "grad_norm": 1.8094528913497925, + "learning_rate": 0.00016443605582246626, + "loss": 0.2727, + "step": 34200 + }, + { + "epoch": 0.5336307481125601, + "grad_norm": 2.224538564682007, + "learning_rate": 0.00016442565670431148, + "loss": 0.3096, + "step": 34210 + }, + { + "epoch": 0.5337867348848817, + "grad_norm": 0.6375226974487305, + "learning_rate": 0.0001644152575861567, + "loss": 0.2251, + "step": 34220 + }, + { + "epoch": 0.5339427216572035, + "grad_norm": 3.727106809616089, + "learning_rate": 0.00016440485846800192, + "loss": 0.4374, + "step": 34230 + }, + { + "epoch": 0.5340987084295252, + "grad_norm": 0.13345426321029663, + "learning_rate": 0.00016439445934984714, + "loss": 0.2011, + "step": 34240 + }, + { + "epoch": 0.5342546952018469, + "grad_norm": 2.1658668518066406, + "learning_rate": 0.00016438406023169236, + "loss": 0.2457, + "step": 34250 + }, + { + "epoch": 0.5344106819741686, + "grad_norm": 19.238407135009766, + "learning_rate": 0.00016437366111353758, + "loss": 0.2756, + "step": 34260 + }, + { + "epoch": 0.5345666687464903, + "grad_norm": 1.0292778015136719, + "learning_rate": 0.0001643632619953828, + "loss": 0.1646, + "step": 34270 + }, + { + "epoch": 0.534722655518812, + "grad_norm": 0.9372987747192383, + "learning_rate": 0.00016435286287722802, + "loss": 0.2762, + "step": 34280 + }, + { + "epoch": 0.5348786422911337, + "grad_norm": 0.3918002247810364, + "learning_rate": 0.00016434246375907324, + "loss": 0.1406, + "step": 34290 + }, + { + "epoch": 0.5350346290634554, + "grad_norm": 1.3518732786178589, + "learning_rate": 0.00016433206464091846, + "loss": 0.2992, + "step": 34300 + }, + { + "epoch": 0.5351906158357771, + "grad_norm": 0.73117595911026, + "learning_rate": 0.00016432166552276368, + "loss": 0.1172, + "step": 34310 + }, + { + "epoch": 0.5353466026080989, + "grad_norm": 0.06655958294868469, + "learning_rate": 0.0001643112664046089, + "loss": 0.1332, + "step": 34320 + }, + { + "epoch": 0.5355025893804205, + "grad_norm": 3.2587168216705322, + "learning_rate": 0.00016430086728645412, + "loss": 0.2052, + "step": 34330 + }, + { + "epoch": 0.5356585761527423, + "grad_norm": 2.6668431758880615, + "learning_rate": 0.00016429046816829934, + "loss": 0.1295, + "step": 34340 + }, + { + "epoch": 0.5358145629250639, + "grad_norm": 1.316080927848816, + "learning_rate": 0.00016428006905014456, + "loss": 0.2072, + "step": 34350 + }, + { + "epoch": 0.5359705496973857, + "grad_norm": 5.144528388977051, + "learning_rate": 0.00016426966993198977, + "loss": 0.1934, + "step": 34360 + }, + { + "epoch": 0.5361265364697073, + "grad_norm": 1.3529599905014038, + "learning_rate": 0.000164259270813835, + "loss": 0.2077, + "step": 34370 + }, + { + "epoch": 0.5362825232420291, + "grad_norm": 0.9132925868034363, + "learning_rate": 0.00016424887169568021, + "loss": 0.2404, + "step": 34380 + }, + { + "epoch": 0.5364385100143508, + "grad_norm": 1.520033597946167, + "learning_rate": 0.00016423847257752543, + "loss": 0.1765, + "step": 34390 + }, + { + "epoch": 0.5365944967866725, + "grad_norm": 2.343975305557251, + "learning_rate": 0.00016422807345937065, + "loss": 0.1959, + "step": 34400 + }, + { + "epoch": 0.5367504835589942, + "grad_norm": 0.5855118632316589, + "learning_rate": 0.00016421767434121587, + "loss": 0.0848, + "step": 34410 + }, + { + "epoch": 0.5369064703313159, + "grad_norm": 1.9108648300170898, + "learning_rate": 0.0001642072752230611, + "loss": 0.2406, + "step": 34420 + }, + { + "epoch": 0.5370624571036376, + "grad_norm": 1.0323792695999146, + "learning_rate": 0.0001641968761049063, + "loss": 0.2298, + "step": 34430 + }, + { + "epoch": 0.5372184438759593, + "grad_norm": 0.17435620725154877, + "learning_rate": 0.00016418647698675153, + "loss": 0.1396, + "step": 34440 + }, + { + "epoch": 0.537374430648281, + "grad_norm": 0.3526577353477478, + "learning_rate": 0.00016417607786859675, + "loss": 0.172, + "step": 34450 + }, + { + "epoch": 0.5375304174206027, + "grad_norm": 2.73685884475708, + "learning_rate": 0.00016416567875044197, + "loss": 0.1167, + "step": 34460 + }, + { + "epoch": 0.5376864041929245, + "grad_norm": 2.765693426132202, + "learning_rate": 0.0001641552796322872, + "loss": 0.1356, + "step": 34470 + }, + { + "epoch": 0.5378423909652461, + "grad_norm": 0.47628021240234375, + "learning_rate": 0.0001641448805141324, + "loss": 0.3104, + "step": 34480 + }, + { + "epoch": 0.5379983777375679, + "grad_norm": 1.1307590007781982, + "learning_rate": 0.00016413448139597763, + "loss": 0.3247, + "step": 34490 + }, + { + "epoch": 0.5381543645098895, + "grad_norm": 0.07136381417512894, + "learning_rate": 0.00016412408227782285, + "loss": 0.319, + "step": 34500 + }, + { + "epoch": 0.5383103512822113, + "grad_norm": 1.7850221395492554, + "learning_rate": 0.00016411368315966807, + "loss": 0.2515, + "step": 34510 + }, + { + "epoch": 0.5384663380545329, + "grad_norm": 1.3870742321014404, + "learning_rate": 0.0001641032840415133, + "loss": 0.256, + "step": 34520 + }, + { + "epoch": 0.5386223248268547, + "grad_norm": 1.1329221725463867, + "learning_rate": 0.0001640928849233585, + "loss": 0.2255, + "step": 34530 + }, + { + "epoch": 0.5387783115991764, + "grad_norm": 0.23236137628555298, + "learning_rate": 0.00016408248580520373, + "loss": 0.4053, + "step": 34540 + }, + { + "epoch": 0.5389342983714981, + "grad_norm": 3.350924253463745, + "learning_rate": 0.00016407208668704895, + "loss": 0.2816, + "step": 34550 + }, + { + "epoch": 0.5390902851438198, + "grad_norm": 1.7459170818328857, + "learning_rate": 0.00016406168756889417, + "loss": 0.0834, + "step": 34560 + }, + { + "epoch": 0.5392462719161415, + "grad_norm": 0.89351487159729, + "learning_rate": 0.0001640512884507394, + "loss": 0.1146, + "step": 34570 + }, + { + "epoch": 0.5394022586884633, + "grad_norm": 0.49410831928253174, + "learning_rate": 0.0001640408893325846, + "loss": 0.2319, + "step": 34580 + }, + { + "epoch": 0.5395582454607849, + "grad_norm": 0.009971237741410732, + "learning_rate": 0.00016403049021442983, + "loss": 0.1674, + "step": 34590 + }, + { + "epoch": 0.5397142322331067, + "grad_norm": 1.420511245727539, + "learning_rate": 0.00016402009109627505, + "loss": 0.1469, + "step": 34600 + }, + { + "epoch": 0.5398702190054283, + "grad_norm": 2.5686769485473633, + "learning_rate": 0.00016400969197812027, + "loss": 0.2607, + "step": 34610 + }, + { + "epoch": 0.5400262057777501, + "grad_norm": 2.5808913707733154, + "learning_rate": 0.00016399929285996549, + "loss": 0.4412, + "step": 34620 + }, + { + "epoch": 0.5401821925500717, + "grad_norm": 0.2778591811656952, + "learning_rate": 0.0001639888937418107, + "loss": 0.1563, + "step": 34630 + }, + { + "epoch": 0.5403381793223935, + "grad_norm": 1.6448099613189697, + "learning_rate": 0.00016397849462365592, + "loss": 0.2484, + "step": 34640 + }, + { + "epoch": 0.5404941660947151, + "grad_norm": 0.003213417250663042, + "learning_rate": 0.00016396809550550114, + "loss": 0.1582, + "step": 34650 + }, + { + "epoch": 0.5406501528670369, + "grad_norm": 2.3433990478515625, + "learning_rate": 0.00016395769638734636, + "loss": 0.2774, + "step": 34660 + }, + { + "epoch": 0.5408061396393585, + "grad_norm": 0.09544426947832108, + "learning_rate": 0.00016394729726919158, + "loss": 0.1242, + "step": 34670 + }, + { + "epoch": 0.5409621264116803, + "grad_norm": 1.8934235572814941, + "learning_rate": 0.0001639368981510368, + "loss": 0.1636, + "step": 34680 + }, + { + "epoch": 0.541118113184002, + "grad_norm": 2.8528554439544678, + "learning_rate": 0.00016392649903288202, + "loss": 0.2524, + "step": 34690 + }, + { + "epoch": 0.5412740999563237, + "grad_norm": 2.837836980819702, + "learning_rate": 0.00016391609991472722, + "loss": 0.1808, + "step": 34700 + }, + { + "epoch": 0.5414300867286455, + "grad_norm": 4.349740028381348, + "learning_rate": 0.00016390570079657246, + "loss": 0.12, + "step": 34710 + }, + { + "epoch": 0.5415860735009671, + "grad_norm": 0.6011945605278015, + "learning_rate": 0.00016389530167841765, + "loss": 0.2865, + "step": 34720 + }, + { + "epoch": 0.5417420602732889, + "grad_norm": 1.2079488039016724, + "learning_rate": 0.0001638849025602629, + "loss": 0.125, + "step": 34730 + }, + { + "epoch": 0.5418980470456105, + "grad_norm": 1.9549164772033691, + "learning_rate": 0.0001638745034421081, + "loss": 0.3285, + "step": 34740 + }, + { + "epoch": 0.5420540338179323, + "grad_norm": 0.6969407200813293, + "learning_rate": 0.00016386410432395334, + "loss": 0.3267, + "step": 34750 + }, + { + "epoch": 0.5422100205902539, + "grad_norm": 1.0561965703964233, + "learning_rate": 0.00016385370520579853, + "loss": 0.1463, + "step": 34760 + }, + { + "epoch": 0.5423660073625757, + "grad_norm": 1.8208931684494019, + "learning_rate": 0.00016384330608764378, + "loss": 0.1713, + "step": 34770 + }, + { + "epoch": 0.5425219941348973, + "grad_norm": 0.6349910497665405, + "learning_rate": 0.000163832906969489, + "loss": 0.2244, + "step": 34780 + }, + { + "epoch": 0.5426779809072191, + "grad_norm": 1.1976515054702759, + "learning_rate": 0.00016382250785133422, + "loss": 0.3407, + "step": 34790 + }, + { + "epoch": 0.5428339676795407, + "grad_norm": 0.5970319509506226, + "learning_rate": 0.00016381210873317944, + "loss": 0.1419, + "step": 34800 + }, + { + "epoch": 0.5429899544518625, + "grad_norm": 0.14623159170150757, + "learning_rate": 0.00016380170961502466, + "loss": 0.14, + "step": 34810 + }, + { + "epoch": 0.5431459412241841, + "grad_norm": 1.4332351684570312, + "learning_rate": 0.00016379131049686988, + "loss": 0.1093, + "step": 34820 + }, + { + "epoch": 0.5433019279965059, + "grad_norm": 1.1443063020706177, + "learning_rate": 0.0001637809113787151, + "loss": 0.3676, + "step": 34830 + }, + { + "epoch": 0.5434579147688277, + "grad_norm": 1.3897461891174316, + "learning_rate": 0.00016377051226056032, + "loss": 0.3713, + "step": 34840 + }, + { + "epoch": 0.5436139015411493, + "grad_norm": 1.7773199081420898, + "learning_rate": 0.00016376011314240554, + "loss": 0.2249, + "step": 34850 + }, + { + "epoch": 0.5437698883134711, + "grad_norm": 0.8524389266967773, + "learning_rate": 0.00016374971402425076, + "loss": 0.2158, + "step": 34860 + }, + { + "epoch": 0.5439258750857927, + "grad_norm": 0.5572389960289001, + "learning_rate": 0.00016373931490609598, + "loss": 0.215, + "step": 34870 + }, + { + "epoch": 0.5440818618581145, + "grad_norm": 3.0238709449768066, + "learning_rate": 0.0001637289157879412, + "loss": 0.3765, + "step": 34880 + }, + { + "epoch": 0.5442378486304361, + "grad_norm": 1.1655421257019043, + "learning_rate": 0.00016371851666978642, + "loss": 0.2491, + "step": 34890 + }, + { + "epoch": 0.5443938354027579, + "grad_norm": 0.49492090940475464, + "learning_rate": 0.00016370811755163164, + "loss": 0.1287, + "step": 34900 + }, + { + "epoch": 0.5445498221750795, + "grad_norm": 0.2732921540737152, + "learning_rate": 0.00016369771843347685, + "loss": 0.2154, + "step": 34910 + }, + { + "epoch": 0.5447058089474013, + "grad_norm": 2.5807769298553467, + "learning_rate": 0.00016368731931532207, + "loss": 0.2032, + "step": 34920 + }, + { + "epoch": 0.5448617957197229, + "grad_norm": 0.5044315457344055, + "learning_rate": 0.0001636769201971673, + "loss": 0.2595, + "step": 34930 + }, + { + "epoch": 0.5450177824920447, + "grad_norm": 2.1248972415924072, + "learning_rate": 0.00016366652107901251, + "loss": 0.5651, + "step": 34940 + }, + { + "epoch": 0.5451737692643663, + "grad_norm": 0.8391468524932861, + "learning_rate": 0.00016365612196085773, + "loss": 0.3028, + "step": 34950 + }, + { + "epoch": 0.5453297560366881, + "grad_norm": 0.36081477999687195, + "learning_rate": 0.00016364572284270295, + "loss": 0.1073, + "step": 34960 + }, + { + "epoch": 0.5454857428090097, + "grad_norm": 3.175804376602173, + "learning_rate": 0.00016363532372454817, + "loss": 0.1889, + "step": 34970 + }, + { + "epoch": 0.5456417295813315, + "grad_norm": 0.20647937059402466, + "learning_rate": 0.0001636249246063934, + "loss": 0.2416, + "step": 34980 + }, + { + "epoch": 0.5457977163536533, + "grad_norm": 2.3525943756103516, + "learning_rate": 0.0001636145254882386, + "loss": 0.3179, + "step": 34990 + }, + { + "epoch": 0.5459537031259749, + "grad_norm": 0.2641216814517975, + "learning_rate": 0.00016360412637008383, + "loss": 0.2157, + "step": 35000 + }, + { + "epoch": 0.5461096898982967, + "grad_norm": 0.09906073659658432, + "learning_rate": 0.00016359372725192905, + "loss": 0.2576, + "step": 35010 + }, + { + "epoch": 0.5462656766706183, + "grad_norm": 0.13971920311450958, + "learning_rate": 0.00016358332813377427, + "loss": 0.1298, + "step": 35020 + }, + { + "epoch": 0.5464216634429401, + "grad_norm": 1.4535638093948364, + "learning_rate": 0.0001635729290156195, + "loss": 0.2357, + "step": 35030 + }, + { + "epoch": 0.5465776502152617, + "grad_norm": 0.9107828736305237, + "learning_rate": 0.0001635625298974647, + "loss": 0.3049, + "step": 35040 + }, + { + "epoch": 0.5467336369875835, + "grad_norm": 0.0377386212348938, + "learning_rate": 0.00016355213077930993, + "loss": 0.2066, + "step": 35050 + }, + { + "epoch": 0.5468896237599051, + "grad_norm": 2.7800869941711426, + "learning_rate": 0.00016354173166115515, + "loss": 0.2974, + "step": 35060 + }, + { + "epoch": 0.5470456105322269, + "grad_norm": 1.6247998476028442, + "learning_rate": 0.00016353133254300037, + "loss": 0.3822, + "step": 35070 + }, + { + "epoch": 0.5472015973045485, + "grad_norm": 1.6479015350341797, + "learning_rate": 0.0001635209334248456, + "loss": 0.3401, + "step": 35080 + }, + { + "epoch": 0.5473575840768703, + "grad_norm": 0.1879737675189972, + "learning_rate": 0.0001635105343066908, + "loss": 0.1549, + "step": 35090 + }, + { + "epoch": 0.5475135708491919, + "grad_norm": 0.39355361461639404, + "learning_rate": 0.00016350013518853603, + "loss": 0.2032, + "step": 35100 + }, + { + "epoch": 0.5476695576215137, + "grad_norm": 2.9976983070373535, + "learning_rate": 0.00016348973607038125, + "loss": 0.3867, + "step": 35110 + }, + { + "epoch": 0.5478255443938354, + "grad_norm": 0.8242707252502441, + "learning_rate": 0.00016347933695222647, + "loss": 0.3601, + "step": 35120 + }, + { + "epoch": 0.5479815311661571, + "grad_norm": 0.874955415725708, + "learning_rate": 0.0001634689378340717, + "loss": 0.2709, + "step": 35130 + }, + { + "epoch": 0.5481375179384789, + "grad_norm": 0.05980971083045006, + "learning_rate": 0.0001634585387159169, + "loss": 0.2646, + "step": 35140 + }, + { + "epoch": 0.5482935047108005, + "grad_norm": 1.8437328338623047, + "learning_rate": 0.0001634481395977621, + "loss": 0.1965, + "step": 35150 + }, + { + "epoch": 0.5484494914831223, + "grad_norm": 0.4142405688762665, + "learning_rate": 0.00016343774047960735, + "loss": 0.1756, + "step": 35160 + }, + { + "epoch": 0.5486054782554439, + "grad_norm": 4.836324214935303, + "learning_rate": 0.00016342734136145254, + "loss": 0.2534, + "step": 35170 + }, + { + "epoch": 0.5487614650277657, + "grad_norm": 1.4933065176010132, + "learning_rate": 0.00016341694224329779, + "loss": 0.2355, + "step": 35180 + }, + { + "epoch": 0.5489174518000873, + "grad_norm": 1.9468894004821777, + "learning_rate": 0.00016340654312514298, + "loss": 0.3113, + "step": 35190 + }, + { + "epoch": 0.5490734385724091, + "grad_norm": 1.012710452079773, + "learning_rate": 0.00016339614400698822, + "loss": 0.2635, + "step": 35200 + }, + { + "epoch": 0.5492294253447307, + "grad_norm": 1.5551140308380127, + "learning_rate": 0.00016338574488883342, + "loss": 0.2951, + "step": 35210 + }, + { + "epoch": 0.5493854121170525, + "grad_norm": 1.8613696098327637, + "learning_rate": 0.00016337534577067866, + "loss": 0.2832, + "step": 35220 + }, + { + "epoch": 0.5495413988893741, + "grad_norm": 1.6748839616775513, + "learning_rate": 0.00016336494665252386, + "loss": 0.2266, + "step": 35230 + }, + { + "epoch": 0.5496973856616959, + "grad_norm": 0.4094032347202301, + "learning_rate": 0.0001633545475343691, + "loss": 0.3497, + "step": 35240 + }, + { + "epoch": 0.5498533724340176, + "grad_norm": 1.6414631605148315, + "learning_rate": 0.0001633441484162143, + "loss": 0.1763, + "step": 35250 + }, + { + "epoch": 0.5500093592063393, + "grad_norm": 1.6380645036697388, + "learning_rate": 0.00016333374929805954, + "loss": 0.1355, + "step": 35260 + }, + { + "epoch": 0.550165345978661, + "grad_norm": 0.7325630187988281, + "learning_rate": 0.00016332335017990473, + "loss": 0.3105, + "step": 35270 + }, + { + "epoch": 0.5503213327509827, + "grad_norm": 1.6015644073486328, + "learning_rate": 0.00016331295106174998, + "loss": 0.2311, + "step": 35280 + }, + { + "epoch": 0.5504773195233045, + "grad_norm": 0.9224210381507874, + "learning_rate": 0.00016330255194359517, + "loss": 0.2746, + "step": 35290 + }, + { + "epoch": 0.5506333062956261, + "grad_norm": 1.30025315284729, + "learning_rate": 0.00016329215282544042, + "loss": 0.2066, + "step": 35300 + }, + { + "epoch": 0.5507892930679479, + "grad_norm": 0.481125146150589, + "learning_rate": 0.0001632817537072856, + "loss": 0.1778, + "step": 35310 + }, + { + "epoch": 0.5509452798402695, + "grad_norm": 0.2709486782550812, + "learning_rate": 0.00016327135458913086, + "loss": 0.1131, + "step": 35320 + }, + { + "epoch": 0.5511012666125913, + "grad_norm": 0.6248563528060913, + "learning_rate": 0.00016326095547097605, + "loss": 0.4356, + "step": 35330 + }, + { + "epoch": 0.5512572533849129, + "grad_norm": 3.0947647094726562, + "learning_rate": 0.0001632505563528213, + "loss": 0.3072, + "step": 35340 + }, + { + "epoch": 0.5514132401572347, + "grad_norm": 1.009535312652588, + "learning_rate": 0.0001632401572346665, + "loss": 0.1808, + "step": 35350 + }, + { + "epoch": 0.5515692269295563, + "grad_norm": 0.042604975402355194, + "learning_rate": 0.00016322975811651174, + "loss": 0.1198, + "step": 35360 + }, + { + "epoch": 0.5517252137018781, + "grad_norm": 1.8652396202087402, + "learning_rate": 0.00016321935899835693, + "loss": 0.2176, + "step": 35370 + }, + { + "epoch": 0.5518812004741998, + "grad_norm": 4.6097187995910645, + "learning_rate": 0.00016320895988020218, + "loss": 0.2132, + "step": 35380 + }, + { + "epoch": 0.5520371872465215, + "grad_norm": 1.3615522384643555, + "learning_rate": 0.00016319856076204737, + "loss": 0.2751, + "step": 35390 + }, + { + "epoch": 0.5521931740188432, + "grad_norm": 0.8844773769378662, + "learning_rate": 0.00016318816164389262, + "loss": 0.22, + "step": 35400 + }, + { + "epoch": 0.5523491607911649, + "grad_norm": 0.5413331985473633, + "learning_rate": 0.0001631777625257378, + "loss": 0.2159, + "step": 35410 + }, + { + "epoch": 0.5525051475634866, + "grad_norm": 1.5456678867340088, + "learning_rate": 0.00016316736340758306, + "loss": 0.2244, + "step": 35420 + }, + { + "epoch": 0.5526611343358083, + "grad_norm": 2.1405861377716064, + "learning_rate": 0.00016315696428942825, + "loss": 0.446, + "step": 35430 + }, + { + "epoch": 0.5528171211081301, + "grad_norm": 1.4269858598709106, + "learning_rate": 0.0001631465651712735, + "loss": 0.3599, + "step": 35440 + }, + { + "epoch": 0.5529731078804517, + "grad_norm": 1.8682516813278198, + "learning_rate": 0.0001631361660531187, + "loss": 0.115, + "step": 35450 + }, + { + "epoch": 0.5531290946527735, + "grad_norm": 1.0175774097442627, + "learning_rate": 0.00016312576693496394, + "loss": 0.4003, + "step": 35460 + }, + { + "epoch": 0.5532850814250951, + "grad_norm": 2.4369170665740967, + "learning_rate": 0.00016311536781680913, + "loss": 0.2489, + "step": 35470 + }, + { + "epoch": 0.5534410681974169, + "grad_norm": 1.886022686958313, + "learning_rate": 0.00016310496869865437, + "loss": 0.4201, + "step": 35480 + }, + { + "epoch": 0.5535970549697385, + "grad_norm": 1.5811959505081177, + "learning_rate": 0.00016309456958049957, + "loss": 0.4286, + "step": 35490 + }, + { + "epoch": 0.5537530417420603, + "grad_norm": 1.971110224723816, + "learning_rate": 0.00016308417046234481, + "loss": 0.3599, + "step": 35500 + }, + { + "epoch": 0.553909028514382, + "grad_norm": 0.0335380844771862, + "learning_rate": 0.00016307377134419, + "loss": 0.3357, + "step": 35510 + }, + { + "epoch": 0.5540650152867037, + "grad_norm": 1.8519577980041504, + "learning_rate": 0.00016306337222603525, + "loss": 0.3126, + "step": 35520 + }, + { + "epoch": 0.5542210020590254, + "grad_norm": 2.1563379764556885, + "learning_rate": 0.00016305297310788045, + "loss": 0.1123, + "step": 35530 + }, + { + "epoch": 0.5543769888313471, + "grad_norm": 1.64332914352417, + "learning_rate": 0.0001630425739897257, + "loss": 0.2853, + "step": 35540 + }, + { + "epoch": 0.5545329756036688, + "grad_norm": 0.061150554567575455, + "learning_rate": 0.00016303217487157088, + "loss": 0.1943, + "step": 35550 + }, + { + "epoch": 0.5546889623759905, + "grad_norm": 1.2701060771942139, + "learning_rate": 0.00016302177575341613, + "loss": 0.2304, + "step": 35560 + }, + { + "epoch": 0.5548449491483122, + "grad_norm": 2.424860715866089, + "learning_rate": 0.00016301137663526132, + "loss": 0.2128, + "step": 35570 + }, + { + "epoch": 0.5550009359206339, + "grad_norm": 0.6803575158119202, + "learning_rate": 0.00016300097751710657, + "loss": 0.1145, + "step": 35580 + }, + { + "epoch": 0.5551569226929557, + "grad_norm": 1.2855092287063599, + "learning_rate": 0.00016299057839895176, + "loss": 0.2448, + "step": 35590 + }, + { + "epoch": 0.5553129094652773, + "grad_norm": 2.6340911388397217, + "learning_rate": 0.00016298017928079698, + "loss": 0.1856, + "step": 35600 + }, + { + "epoch": 0.5554688962375991, + "grad_norm": 1.776382327079773, + "learning_rate": 0.0001629697801626422, + "loss": 0.2422, + "step": 35610 + }, + { + "epoch": 0.5556248830099207, + "grad_norm": 3.0746347904205322, + "learning_rate": 0.00016295938104448742, + "loss": 0.1578, + "step": 35620 + }, + { + "epoch": 0.5557808697822425, + "grad_norm": 1.4887659549713135, + "learning_rate": 0.00016294898192633267, + "loss": 0.2385, + "step": 35630 + }, + { + "epoch": 0.5559368565545642, + "grad_norm": 0.5371220707893372, + "learning_rate": 0.00016293858280817786, + "loss": 0.2361, + "step": 35640 + }, + { + "epoch": 0.5560928433268859, + "grad_norm": 0.18604904413223267, + "learning_rate": 0.0001629281836900231, + "loss": 0.0828, + "step": 35650 + }, + { + "epoch": 0.5562488300992076, + "grad_norm": 2.267854690551758, + "learning_rate": 0.0001629177845718683, + "loss": 0.413, + "step": 35660 + }, + { + "epoch": 0.5564048168715293, + "grad_norm": 1.9755452871322632, + "learning_rate": 0.00016290738545371355, + "loss": 0.146, + "step": 35670 + }, + { + "epoch": 0.556560803643851, + "grad_norm": 2.430293321609497, + "learning_rate": 0.00016289698633555874, + "loss": 0.2677, + "step": 35680 + }, + { + "epoch": 0.5567167904161727, + "grad_norm": 0.6319543719291687, + "learning_rate": 0.000162886587217404, + "loss": 0.1481, + "step": 35690 + }, + { + "epoch": 0.5568727771884944, + "grad_norm": 0.17439277470111847, + "learning_rate": 0.00016287618809924918, + "loss": 0.1641, + "step": 35700 + }, + { + "epoch": 0.5570287639608161, + "grad_norm": 0.350175142288208, + "learning_rate": 0.00016286578898109443, + "loss": 0.393, + "step": 35710 + }, + { + "epoch": 0.5571847507331378, + "grad_norm": 2.641941547393799, + "learning_rate": 0.00016285538986293962, + "loss": 0.1907, + "step": 35720 + }, + { + "epoch": 0.5573407375054595, + "grad_norm": 0.1910303384065628, + "learning_rate": 0.00016284499074478487, + "loss": 0.2248, + "step": 35730 + }, + { + "epoch": 0.5574967242777813, + "grad_norm": 1.0180896520614624, + "learning_rate": 0.00016283459162663006, + "loss": 0.1924, + "step": 35740 + }, + { + "epoch": 0.557652711050103, + "grad_norm": 1.2047260999679565, + "learning_rate": 0.0001628241925084753, + "loss": 0.2031, + "step": 35750 + }, + { + "epoch": 0.5578086978224247, + "grad_norm": 1.8182405233383179, + "learning_rate": 0.0001628137933903205, + "loss": 0.1573, + "step": 35760 + }, + { + "epoch": 0.5579646845947464, + "grad_norm": 0.5485963821411133, + "learning_rate": 0.00016280339427216574, + "loss": 0.152, + "step": 35770 + }, + { + "epoch": 0.5581206713670681, + "grad_norm": 0.1674145758152008, + "learning_rate": 0.00016279299515401094, + "loss": 0.2388, + "step": 35780 + }, + { + "epoch": 0.5582766581393898, + "grad_norm": 0.36187657713890076, + "learning_rate": 0.00016278259603585618, + "loss": 0.1618, + "step": 35790 + }, + { + "epoch": 0.5584326449117115, + "grad_norm": 0.013768521137535572, + "learning_rate": 0.00016277219691770138, + "loss": 0.3666, + "step": 35800 + }, + { + "epoch": 0.5585886316840332, + "grad_norm": 2.8602802753448486, + "learning_rate": 0.00016276179779954662, + "loss": 0.2065, + "step": 35810 + }, + { + "epoch": 0.5587446184563549, + "grad_norm": 1.7925455570220947, + "learning_rate": 0.00016275139868139182, + "loss": 0.1729, + "step": 35820 + }, + { + "epoch": 0.5589006052286766, + "grad_norm": 2.910456418991089, + "learning_rate": 0.00016274099956323706, + "loss": 0.3734, + "step": 35830 + }, + { + "epoch": 0.5590565920009983, + "grad_norm": 0.8378308415412903, + "learning_rate": 0.00016273060044508225, + "loss": 0.2255, + "step": 35840 + }, + { + "epoch": 0.55921257877332, + "grad_norm": 0.409534752368927, + "learning_rate": 0.0001627202013269275, + "loss": 0.2046, + "step": 35850 + }, + { + "epoch": 0.5593685655456417, + "grad_norm": 0.49498099088668823, + "learning_rate": 0.0001627098022087727, + "loss": 0.1058, + "step": 35860 + }, + { + "epoch": 0.5595245523179634, + "grad_norm": 1.1761783361434937, + "learning_rate": 0.00016269940309061794, + "loss": 0.1566, + "step": 35870 + }, + { + "epoch": 0.5596805390902851, + "grad_norm": 3.8204751014709473, + "learning_rate": 0.00016268900397246313, + "loss": 0.2646, + "step": 35880 + }, + { + "epoch": 0.5598365258626069, + "grad_norm": 0.9882522225379944, + "learning_rate": 0.00016267860485430838, + "loss": 0.1756, + "step": 35890 + }, + { + "epoch": 0.5599925126349286, + "grad_norm": 1.1832259893417358, + "learning_rate": 0.00016266820573615357, + "loss": 0.1385, + "step": 35900 + }, + { + "epoch": 0.5601484994072503, + "grad_norm": 0.7638296484947205, + "learning_rate": 0.00016265780661799882, + "loss": 0.5679, + "step": 35910 + }, + { + "epoch": 0.560304486179572, + "grad_norm": 0.8551504611968994, + "learning_rate": 0.000162647407499844, + "loss": 0.2981, + "step": 35920 + }, + { + "epoch": 0.5604604729518937, + "grad_norm": 5.499948501586914, + "learning_rate": 0.00016263700838168926, + "loss": 0.1429, + "step": 35930 + }, + { + "epoch": 0.5606164597242154, + "grad_norm": 3.0494496822357178, + "learning_rate": 0.00016262660926353445, + "loss": 0.1322, + "step": 35940 + }, + { + "epoch": 0.5607724464965371, + "grad_norm": 0.2482384592294693, + "learning_rate": 0.0001626162101453797, + "loss": 0.2473, + "step": 35950 + }, + { + "epoch": 0.5609284332688588, + "grad_norm": 0.009750776924192905, + "learning_rate": 0.0001626058110272249, + "loss": 0.0429, + "step": 35960 + }, + { + "epoch": 0.5610844200411805, + "grad_norm": 0.5742604732513428, + "learning_rate": 0.00016259541190907014, + "loss": 0.4565, + "step": 35970 + }, + { + "epoch": 0.5612404068135022, + "grad_norm": 0.7810243368148804, + "learning_rate": 0.00016258501279091533, + "loss": 0.2107, + "step": 35980 + }, + { + "epoch": 0.5613963935858239, + "grad_norm": 2.785747528076172, + "learning_rate": 0.00016257461367276058, + "loss": 0.5087, + "step": 35990 + }, + { + "epoch": 0.5615523803581456, + "grad_norm": 0.8806902766227722, + "learning_rate": 0.00016256421455460577, + "loss": 0.2937, + "step": 36000 + }, + { + "epoch": 0.5617083671304673, + "grad_norm": 1.857373833656311, + "learning_rate": 0.00016255381543645102, + "loss": 0.5042, + "step": 36010 + }, + { + "epoch": 0.561864353902789, + "grad_norm": 0.5333901047706604, + "learning_rate": 0.0001625434163182962, + "loss": 0.1678, + "step": 36020 + }, + { + "epoch": 0.5620203406751108, + "grad_norm": 11.023160934448242, + "learning_rate": 0.00016253301720014145, + "loss": 0.2721, + "step": 36030 + }, + { + "epoch": 0.5621763274474325, + "grad_norm": 2.6214029788970947, + "learning_rate": 0.00016252261808198665, + "loss": 0.1675, + "step": 36040 + }, + { + "epoch": 0.5623323142197542, + "grad_norm": 0.25980064272880554, + "learning_rate": 0.0001625122189638319, + "loss": 0.1832, + "step": 36050 + }, + { + "epoch": 0.5624883009920759, + "grad_norm": 1.3559473752975464, + "learning_rate": 0.0001625018198456771, + "loss": 0.121, + "step": 36060 + }, + { + "epoch": 0.5626442877643976, + "grad_norm": 0.3558153212070465, + "learning_rate": 0.0001624914207275223, + "loss": 0.2797, + "step": 36070 + }, + { + "epoch": 0.5628002745367193, + "grad_norm": 0.6314427256584167, + "learning_rate": 0.00016248102160936753, + "loss": 0.3455, + "step": 36080 + }, + { + "epoch": 0.562956261309041, + "grad_norm": 0.16120101511478424, + "learning_rate": 0.00016247062249121275, + "loss": 0.3669, + "step": 36090 + }, + { + "epoch": 0.5631122480813627, + "grad_norm": 1.041572093963623, + "learning_rate": 0.00016246022337305797, + "loss": 0.2744, + "step": 36100 + }, + { + "epoch": 0.5632682348536844, + "grad_norm": 0.2645890414714813, + "learning_rate": 0.00016244982425490318, + "loss": 0.1487, + "step": 36110 + }, + { + "epoch": 0.5634242216260061, + "grad_norm": 5.698698043823242, + "learning_rate": 0.0001624394251367484, + "loss": 0.173, + "step": 36120 + }, + { + "epoch": 0.5635802083983278, + "grad_norm": 0.3299804627895355, + "learning_rate": 0.00016242902601859362, + "loss": 0.1965, + "step": 36130 + }, + { + "epoch": 0.5637361951706495, + "grad_norm": 1.2476481199264526, + "learning_rate": 0.00016241862690043884, + "loss": 0.4097, + "step": 36140 + }, + { + "epoch": 0.5638921819429712, + "grad_norm": 0.30926263332366943, + "learning_rate": 0.00016240822778228406, + "loss": 0.1416, + "step": 36150 + }, + { + "epoch": 0.564048168715293, + "grad_norm": 0.05872740224003792, + "learning_rate": 0.00016239782866412928, + "loss": 0.1449, + "step": 36160 + }, + { + "epoch": 0.5642041554876146, + "grad_norm": 0.2292211800813675, + "learning_rate": 0.0001623874295459745, + "loss": 0.2557, + "step": 36170 + }, + { + "epoch": 0.5643601422599364, + "grad_norm": 1.7822531461715698, + "learning_rate": 0.00016237703042781972, + "loss": 0.2953, + "step": 36180 + }, + { + "epoch": 0.5645161290322581, + "grad_norm": 2.4908461570739746, + "learning_rate": 0.00016236663130966494, + "loss": 0.3888, + "step": 36190 + }, + { + "epoch": 0.5646721158045798, + "grad_norm": 0.08363594114780426, + "learning_rate": 0.00016235623219151016, + "loss": 0.2293, + "step": 36200 + }, + { + "epoch": 0.5648281025769015, + "grad_norm": 1.4830002784729004, + "learning_rate": 0.00016234583307335538, + "loss": 0.2348, + "step": 36210 + }, + { + "epoch": 0.5649840893492232, + "grad_norm": 1.8443071842193604, + "learning_rate": 0.0001623354339552006, + "loss": 0.3392, + "step": 36220 + }, + { + "epoch": 0.5651400761215449, + "grad_norm": 2.8050875663757324, + "learning_rate": 0.00016232503483704582, + "loss": 0.2426, + "step": 36230 + }, + { + "epoch": 0.5652960628938666, + "grad_norm": 3.6627259254455566, + "learning_rate": 0.00016231463571889104, + "loss": 0.2118, + "step": 36240 + }, + { + "epoch": 0.5654520496661883, + "grad_norm": 0.3443094789981842, + "learning_rate": 0.00016230423660073626, + "loss": 0.2418, + "step": 36250 + }, + { + "epoch": 0.56560803643851, + "grad_norm": 1.7411112785339355, + "learning_rate": 0.00016229383748258148, + "loss": 0.1628, + "step": 36260 + }, + { + "epoch": 0.5657640232108317, + "grad_norm": 1.1321816444396973, + "learning_rate": 0.0001622834383644267, + "loss": 0.2732, + "step": 36270 + }, + { + "epoch": 0.5659200099831534, + "grad_norm": 2.9686615467071533, + "learning_rate": 0.00016227303924627192, + "loss": 0.3036, + "step": 36280 + }, + { + "epoch": 0.5660759967554752, + "grad_norm": 5.03530740737915, + "learning_rate": 0.00016226264012811714, + "loss": 0.2143, + "step": 36290 + }, + { + "epoch": 0.5662319835277968, + "grad_norm": 2.3943281173706055, + "learning_rate": 0.00016225224100996236, + "loss": 0.2629, + "step": 36300 + }, + { + "epoch": 0.5663879703001186, + "grad_norm": 1.812828779220581, + "learning_rate": 0.00016224184189180758, + "loss": 0.3892, + "step": 36310 + }, + { + "epoch": 0.5665439570724402, + "grad_norm": 1.4708483219146729, + "learning_rate": 0.0001622314427736528, + "loss": 0.2118, + "step": 36320 + }, + { + "epoch": 0.566699943844762, + "grad_norm": 8.99713134765625, + "learning_rate": 0.00016222104365549802, + "loss": 0.3185, + "step": 36330 + }, + { + "epoch": 0.5668559306170837, + "grad_norm": 1.7472341060638428, + "learning_rate": 0.00016221064453734324, + "loss": 0.1659, + "step": 36340 + }, + { + "epoch": 0.5670119173894054, + "grad_norm": 0.8059778213500977, + "learning_rate": 0.00016220024541918846, + "loss": 0.0621, + "step": 36350 + }, + { + "epoch": 0.5671679041617271, + "grad_norm": 4.0501861572265625, + "learning_rate": 0.00016218984630103368, + "loss": 0.1399, + "step": 36360 + }, + { + "epoch": 0.5673238909340488, + "grad_norm": 0.014400321058928967, + "learning_rate": 0.0001621794471828789, + "loss": 0.2372, + "step": 36370 + }, + { + "epoch": 0.5674798777063705, + "grad_norm": 0.4061933755874634, + "learning_rate": 0.00016216904806472412, + "loss": 0.3854, + "step": 36380 + }, + { + "epoch": 0.5676358644786922, + "grad_norm": 0.4940955340862274, + "learning_rate": 0.00016215864894656933, + "loss": 0.1689, + "step": 36390 + }, + { + "epoch": 0.567791851251014, + "grad_norm": 0.3220517635345459, + "learning_rate": 0.00016214824982841455, + "loss": 0.1912, + "step": 36400 + }, + { + "epoch": 0.5679478380233356, + "grad_norm": 2.0848920345306396, + "learning_rate": 0.00016213785071025977, + "loss": 0.1639, + "step": 36410 + }, + { + "epoch": 0.5681038247956574, + "grad_norm": 0.04129406809806824, + "learning_rate": 0.000162127451592105, + "loss": 0.1378, + "step": 36420 + }, + { + "epoch": 0.568259811567979, + "grad_norm": 1.9343401193618774, + "learning_rate": 0.0001621170524739502, + "loss": 0.3759, + "step": 36430 + }, + { + "epoch": 0.5684157983403008, + "grad_norm": 1.0625615119934082, + "learning_rate": 0.00016210665335579543, + "loss": 0.2074, + "step": 36440 + }, + { + "epoch": 0.5685717851126224, + "grad_norm": 0.14784491062164307, + "learning_rate": 0.00016209625423764065, + "loss": 0.0981, + "step": 36450 + }, + { + "epoch": 0.5687277718849442, + "grad_norm": 1.8117821216583252, + "learning_rate": 0.00016208585511948587, + "loss": 0.2755, + "step": 36460 + }, + { + "epoch": 0.5688837586572658, + "grad_norm": 0.061010006815195084, + "learning_rate": 0.0001620754560013311, + "loss": 0.1815, + "step": 36470 + }, + { + "epoch": 0.5690397454295876, + "grad_norm": 0.44453224539756775, + "learning_rate": 0.0001620650568831763, + "loss": 0.0891, + "step": 36480 + }, + { + "epoch": 0.5691957322019093, + "grad_norm": 0.6130645871162415, + "learning_rate": 0.00016205465776502153, + "loss": 0.2778, + "step": 36490 + }, + { + "epoch": 0.569351718974231, + "grad_norm": 5.408796787261963, + "learning_rate": 0.00016204425864686678, + "loss": 0.2301, + "step": 36500 + }, + { + "epoch": 0.5695077057465527, + "grad_norm": 1.2095412015914917, + "learning_rate": 0.00016203385952871197, + "loss": 0.1378, + "step": 36510 + }, + { + "epoch": 0.5696636925188744, + "grad_norm": 0.46873530745506287, + "learning_rate": 0.0001620234604105572, + "loss": 0.1431, + "step": 36520 + }, + { + "epoch": 0.5698196792911961, + "grad_norm": 1.028438925743103, + "learning_rate": 0.0001620130612924024, + "loss": 0.1114, + "step": 36530 + }, + { + "epoch": 0.5699756660635178, + "grad_norm": 0.5952587127685547, + "learning_rate": 0.00016200266217424763, + "loss": 0.2156, + "step": 36540 + }, + { + "epoch": 0.5701316528358396, + "grad_norm": 0.5134285688400269, + "learning_rate": 0.00016199226305609285, + "loss": 0.2712, + "step": 36550 + }, + { + "epoch": 0.5702876396081612, + "grad_norm": 0.03440163657069206, + "learning_rate": 0.00016198186393793807, + "loss": 0.312, + "step": 36560 + }, + { + "epoch": 0.570443626380483, + "grad_norm": 0.10363951325416565, + "learning_rate": 0.0001619714648197833, + "loss": 0.2399, + "step": 36570 + }, + { + "epoch": 0.5705996131528046, + "grad_norm": 1.3341947793960571, + "learning_rate": 0.0001619610657016285, + "loss": 0.1397, + "step": 36580 + }, + { + "epoch": 0.5707555999251264, + "grad_norm": 0.0668073296546936, + "learning_rate": 0.00016195066658347373, + "loss": 0.1091, + "step": 36590 + }, + { + "epoch": 0.570911586697448, + "grad_norm": 5.0486674308776855, + "learning_rate": 0.00016194026746531895, + "loss": 0.4893, + "step": 36600 + }, + { + "epoch": 0.5710675734697698, + "grad_norm": 1.2655692100524902, + "learning_rate": 0.00016192986834716417, + "loss": 0.3344, + "step": 36610 + }, + { + "epoch": 0.5712235602420914, + "grad_norm": 0.1458190381526947, + "learning_rate": 0.00016191946922900939, + "loss": 0.247, + "step": 36620 + }, + { + "epoch": 0.5713795470144132, + "grad_norm": 2.5789072513580322, + "learning_rate": 0.0001619090701108546, + "loss": 0.1546, + "step": 36630 + }, + { + "epoch": 0.5715355337867349, + "grad_norm": 1.191821575164795, + "learning_rate": 0.00016189867099269983, + "loss": 0.2215, + "step": 36640 + }, + { + "epoch": 0.5716915205590566, + "grad_norm": 8.163617134094238, + "learning_rate": 0.00016188827187454505, + "loss": 0.1696, + "step": 36650 + }, + { + "epoch": 0.5718475073313783, + "grad_norm": 3.162277936935425, + "learning_rate": 0.00016187787275639027, + "loss": 0.2906, + "step": 36660 + }, + { + "epoch": 0.5720034941037, + "grad_norm": 2.2368602752685547, + "learning_rate": 0.00016186747363823548, + "loss": 0.3806, + "step": 36670 + }, + { + "epoch": 0.5721594808760218, + "grad_norm": 2.0193803310394287, + "learning_rate": 0.0001618570745200807, + "loss": 0.1964, + "step": 36680 + }, + { + "epoch": 0.5723154676483434, + "grad_norm": 0.9131811261177063, + "learning_rate": 0.00016184667540192592, + "loss": 0.1902, + "step": 36690 + }, + { + "epoch": 0.5724714544206652, + "grad_norm": 0.21330870687961578, + "learning_rate": 0.00016183627628377114, + "loss": 0.2907, + "step": 36700 + }, + { + "epoch": 0.5726274411929868, + "grad_norm": 0.7145895957946777, + "learning_rate": 0.00016182587716561636, + "loss": 0.2574, + "step": 36710 + }, + { + "epoch": 0.5727834279653086, + "grad_norm": 1.0699433088302612, + "learning_rate": 0.00016181547804746158, + "loss": 0.1675, + "step": 36720 + }, + { + "epoch": 0.5729394147376302, + "grad_norm": 0.5066679120063782, + "learning_rate": 0.0001618050789293068, + "loss": 0.4487, + "step": 36730 + }, + { + "epoch": 0.573095401509952, + "grad_norm": 0.7482333183288574, + "learning_rate": 0.00016179467981115202, + "loss": 0.4101, + "step": 36740 + }, + { + "epoch": 0.5732513882822736, + "grad_norm": 0.15925532579421997, + "learning_rate": 0.00016178428069299724, + "loss": 0.2825, + "step": 36750 + }, + { + "epoch": 0.5734073750545954, + "grad_norm": 0.95689457654953, + "learning_rate": 0.00016177388157484246, + "loss": 0.2754, + "step": 36760 + }, + { + "epoch": 0.573563361826917, + "grad_norm": 2.15043568611145, + "learning_rate": 0.00016176348245668768, + "loss": 0.2073, + "step": 36770 + }, + { + "epoch": 0.5737193485992388, + "grad_norm": 2.651520252227783, + "learning_rate": 0.0001617530833385329, + "loss": 0.2517, + "step": 36780 + }, + { + "epoch": 0.5738753353715605, + "grad_norm": 2.1156702041625977, + "learning_rate": 0.00016174268422037812, + "loss": 0.3029, + "step": 36790 + }, + { + "epoch": 0.5740313221438822, + "grad_norm": 3.6200709342956543, + "learning_rate": 0.00016173228510222334, + "loss": 0.4174, + "step": 36800 + }, + { + "epoch": 0.574187308916204, + "grad_norm": 1.447937250137329, + "learning_rate": 0.00016172188598406856, + "loss": 0.1763, + "step": 36810 + }, + { + "epoch": 0.5743432956885256, + "grad_norm": 1.4020819664001465, + "learning_rate": 0.00016171148686591378, + "loss": 0.3994, + "step": 36820 + }, + { + "epoch": 0.5744992824608474, + "grad_norm": 1.4079807996749878, + "learning_rate": 0.000161701087747759, + "loss": 0.2051, + "step": 36830 + }, + { + "epoch": 0.574655269233169, + "grad_norm": 0.05268567427992821, + "learning_rate": 0.00016169068862960422, + "loss": 0.1773, + "step": 36840 + }, + { + "epoch": 0.5748112560054908, + "grad_norm": 2.248767614364624, + "learning_rate": 0.00016168028951144944, + "loss": 0.2534, + "step": 36850 + }, + { + "epoch": 0.5749672427778124, + "grad_norm": 0.3726302981376648, + "learning_rate": 0.00016166989039329466, + "loss": 0.328, + "step": 36860 + }, + { + "epoch": 0.5751232295501342, + "grad_norm": 1.0772110223770142, + "learning_rate": 0.00016165949127513988, + "loss": 0.2069, + "step": 36870 + }, + { + "epoch": 0.5752792163224558, + "grad_norm": 0.11339398473501205, + "learning_rate": 0.0001616490921569851, + "loss": 0.1827, + "step": 36880 + }, + { + "epoch": 0.5754352030947776, + "grad_norm": 1.2981103658676147, + "learning_rate": 0.00016163869303883032, + "loss": 0.1378, + "step": 36890 + }, + { + "epoch": 0.5755911898670992, + "grad_norm": 0.3302285075187683, + "learning_rate": 0.00016162829392067554, + "loss": 0.2311, + "step": 36900 + }, + { + "epoch": 0.575747176639421, + "grad_norm": 0.10168436169624329, + "learning_rate": 0.00016161789480252076, + "loss": 0.1678, + "step": 36910 + }, + { + "epoch": 0.5759031634117426, + "grad_norm": 1.6778357028961182, + "learning_rate": 0.00016160749568436598, + "loss": 0.2385, + "step": 36920 + }, + { + "epoch": 0.5760591501840644, + "grad_norm": 8.1710205078125, + "learning_rate": 0.0001615970965662112, + "loss": 0.3287, + "step": 36930 + }, + { + "epoch": 0.5762151369563862, + "grad_norm": 2.6481053829193115, + "learning_rate": 0.00016158669744805641, + "loss": 0.3119, + "step": 36940 + }, + { + "epoch": 0.5763711237287078, + "grad_norm": 0.3120105266571045, + "learning_rate": 0.00016157629832990163, + "loss": 0.1985, + "step": 36950 + }, + { + "epoch": 0.5765271105010296, + "grad_norm": 2.7221410274505615, + "learning_rate": 0.00016156589921174685, + "loss": 0.3697, + "step": 36960 + }, + { + "epoch": 0.5766830972733512, + "grad_norm": 0.34397706389427185, + "learning_rate": 0.00016155550009359207, + "loss": 0.2162, + "step": 36970 + }, + { + "epoch": 0.576839084045673, + "grad_norm": 4.6421003341674805, + "learning_rate": 0.0001615451009754373, + "loss": 0.2892, + "step": 36980 + }, + { + "epoch": 0.5769950708179946, + "grad_norm": 1.5395855903625488, + "learning_rate": 0.0001615347018572825, + "loss": 0.1258, + "step": 36990 + }, + { + "epoch": 0.5771510575903164, + "grad_norm": 0.4586782455444336, + "learning_rate": 0.00016152430273912773, + "loss": 0.1752, + "step": 37000 + }, + { + "epoch": 0.577307044362638, + "grad_norm": 0.4277658760547638, + "learning_rate": 0.00016151390362097295, + "loss": 0.1722, + "step": 37010 + }, + { + "epoch": 0.5774630311349598, + "grad_norm": 1.0282249450683594, + "learning_rate": 0.00016150350450281817, + "loss": 0.1361, + "step": 37020 + }, + { + "epoch": 0.5776190179072814, + "grad_norm": 0.2496921867132187, + "learning_rate": 0.0001614931053846634, + "loss": 0.0954, + "step": 37030 + }, + { + "epoch": 0.5777750046796032, + "grad_norm": 1.7900562286376953, + "learning_rate": 0.0001614827062665086, + "loss": 0.1284, + "step": 37040 + }, + { + "epoch": 0.5779309914519248, + "grad_norm": 0.6629091501235962, + "learning_rate": 0.00016147230714835383, + "loss": 0.2657, + "step": 37050 + }, + { + "epoch": 0.5780869782242466, + "grad_norm": 2.502882242202759, + "learning_rate": 0.00016146190803019905, + "loss": 0.207, + "step": 37060 + }, + { + "epoch": 0.5782429649965682, + "grad_norm": 1.486069917678833, + "learning_rate": 0.00016145150891204427, + "loss": 0.1232, + "step": 37070 + }, + { + "epoch": 0.57839895176889, + "grad_norm": 0.13543163239955902, + "learning_rate": 0.0001614411097938895, + "loss": 0.3245, + "step": 37080 + }, + { + "epoch": 0.5785549385412118, + "grad_norm": 3.9703898429870605, + "learning_rate": 0.0001614307106757347, + "loss": 0.3104, + "step": 37090 + }, + { + "epoch": 0.5787109253135334, + "grad_norm": 3.2236313819885254, + "learning_rate": 0.00016142031155757993, + "loss": 0.1739, + "step": 37100 + }, + { + "epoch": 0.5788669120858552, + "grad_norm": 0.3854866325855255, + "learning_rate": 0.00016140991243942515, + "loss": 0.084, + "step": 37110 + }, + { + "epoch": 0.5790228988581768, + "grad_norm": 0.1981169581413269, + "learning_rate": 0.00016139951332127037, + "loss": 0.2367, + "step": 37120 + }, + { + "epoch": 0.5791788856304986, + "grad_norm": 2.7315785884857178, + "learning_rate": 0.0001613891142031156, + "loss": 0.1816, + "step": 37130 + }, + { + "epoch": 0.5793348724028202, + "grad_norm": 0.48251378536224365, + "learning_rate": 0.0001613787150849608, + "loss": 0.2123, + "step": 37140 + }, + { + "epoch": 0.579490859175142, + "grad_norm": 5.0431413650512695, + "learning_rate": 0.00016136831596680603, + "loss": 0.2489, + "step": 37150 + }, + { + "epoch": 0.5796468459474636, + "grad_norm": 0.81981360912323, + "learning_rate": 0.00016135791684865125, + "loss": 0.5235, + "step": 37160 + }, + { + "epoch": 0.5798028327197854, + "grad_norm": 0.5701965689659119, + "learning_rate": 0.00016134751773049647, + "loss": 0.1709, + "step": 37170 + }, + { + "epoch": 0.579958819492107, + "grad_norm": 1.3905616998672485, + "learning_rate": 0.00016133711861234169, + "loss": 0.2641, + "step": 37180 + }, + { + "epoch": 0.5801148062644288, + "grad_norm": 2.281285047531128, + "learning_rate": 0.0001613267194941869, + "loss": 0.3915, + "step": 37190 + }, + { + "epoch": 0.5802707930367504, + "grad_norm": 1.2038309574127197, + "learning_rate": 0.00016131632037603213, + "loss": 0.0881, + "step": 37200 + }, + { + "epoch": 0.5804267798090722, + "grad_norm": 0.3096400201320648, + "learning_rate": 0.00016130592125787735, + "loss": 0.1341, + "step": 37210 + }, + { + "epoch": 0.5805827665813938, + "grad_norm": 0.10219905525445938, + "learning_rate": 0.00016129552213972256, + "loss": 0.0867, + "step": 37220 + }, + { + "epoch": 0.5807387533537156, + "grad_norm": 0.13869646191596985, + "learning_rate": 0.00016128512302156778, + "loss": 0.1621, + "step": 37230 + }, + { + "epoch": 0.5808947401260373, + "grad_norm": 0.7992938160896301, + "learning_rate": 0.000161274723903413, + "loss": 0.275, + "step": 37240 + }, + { + "epoch": 0.581050726898359, + "grad_norm": 0.2635972201824188, + "learning_rate": 0.00016126432478525822, + "loss": 0.2429, + "step": 37250 + }, + { + "epoch": 0.5812067136706808, + "grad_norm": 1.7211238145828247, + "learning_rate": 0.00016125392566710344, + "loss": 0.1541, + "step": 37260 + }, + { + "epoch": 0.5813627004430024, + "grad_norm": 2.2448506355285645, + "learning_rate": 0.00016124352654894866, + "loss": 0.3689, + "step": 37270 + }, + { + "epoch": 0.5815186872153242, + "grad_norm": 1.238708257675171, + "learning_rate": 0.00016123312743079388, + "loss": 0.3292, + "step": 37280 + }, + { + "epoch": 0.5816746739876458, + "grad_norm": 0.5670503377914429, + "learning_rate": 0.0001612227283126391, + "loss": 0.1151, + "step": 37290 + }, + { + "epoch": 0.5818306607599676, + "grad_norm": 2.287304401397705, + "learning_rate": 0.00016121232919448432, + "loss": 0.3558, + "step": 37300 + }, + { + "epoch": 0.5819866475322892, + "grad_norm": 0.8009629845619202, + "learning_rate": 0.00016120193007632954, + "loss": 0.1861, + "step": 37310 + }, + { + "epoch": 0.582142634304611, + "grad_norm": 1.1901466846466064, + "learning_rate": 0.00016119153095817476, + "loss": 0.089, + "step": 37320 + }, + { + "epoch": 0.5822986210769326, + "grad_norm": 1.7198141813278198, + "learning_rate": 0.00016118113184001998, + "loss": 0.2931, + "step": 37330 + }, + { + "epoch": 0.5824546078492544, + "grad_norm": 0.8832749724388123, + "learning_rate": 0.0001611707327218652, + "loss": 0.2342, + "step": 37340 + }, + { + "epoch": 0.582610594621576, + "grad_norm": 1.0623290538787842, + "learning_rate": 0.00016116033360371042, + "loss": 0.2317, + "step": 37350 + }, + { + "epoch": 0.5827665813938978, + "grad_norm": 0.5173699855804443, + "learning_rate": 0.00016114993448555564, + "loss": 0.1299, + "step": 37360 + }, + { + "epoch": 0.5829225681662195, + "grad_norm": 0.05118720978498459, + "learning_rate": 0.00016113953536740086, + "loss": 0.0786, + "step": 37370 + }, + { + "epoch": 0.5830785549385412, + "grad_norm": 2.0060107707977295, + "learning_rate": 0.00016112913624924608, + "loss": 0.266, + "step": 37380 + }, + { + "epoch": 0.5832345417108629, + "grad_norm": 5.121693134307861, + "learning_rate": 0.0001611187371310913, + "loss": 0.1824, + "step": 37390 + }, + { + "epoch": 0.5833905284831846, + "grad_norm": 0.3568089008331299, + "learning_rate": 0.00016110833801293652, + "loss": 0.3189, + "step": 37400 + }, + { + "epoch": 0.5835465152555064, + "grad_norm": 0.8631492853164673, + "learning_rate": 0.00016109793889478174, + "loss": 0.1724, + "step": 37410 + }, + { + "epoch": 0.583702502027828, + "grad_norm": 1.9214404821395874, + "learning_rate": 0.00016108753977662693, + "loss": 0.2094, + "step": 37420 + }, + { + "epoch": 0.5838584888001498, + "grad_norm": 1.9055646657943726, + "learning_rate": 0.00016107714065847218, + "loss": 0.1258, + "step": 37430 + }, + { + "epoch": 0.5840144755724714, + "grad_norm": 0.6614194512367249, + "learning_rate": 0.00016106674154031737, + "loss": 0.1966, + "step": 37440 + }, + { + "epoch": 0.5841704623447932, + "grad_norm": 0.286883682012558, + "learning_rate": 0.00016105634242216262, + "loss": 0.588, + "step": 37450 + }, + { + "epoch": 0.5843264491171148, + "grad_norm": 0.5599405765533447, + "learning_rate": 0.0001610459433040078, + "loss": 0.3414, + "step": 37460 + }, + { + "epoch": 0.5844824358894366, + "grad_norm": 1.5908915996551514, + "learning_rate": 0.00016103554418585306, + "loss": 0.1264, + "step": 37470 + }, + { + "epoch": 0.5846384226617583, + "grad_norm": 1.0729930400848389, + "learning_rate": 0.00016102514506769825, + "loss": 0.4815, + "step": 37480 + }, + { + "epoch": 0.58479440943408, + "grad_norm": 0.8912356495857239, + "learning_rate": 0.0001610147459495435, + "loss": 0.2092, + "step": 37490 + }, + { + "epoch": 0.5849503962064017, + "grad_norm": 0.24116197228431702, + "learning_rate": 0.0001610043468313887, + "loss": 0.1552, + "step": 37500 + }, + { + "epoch": 0.5851063829787234, + "grad_norm": 2.245192527770996, + "learning_rate": 0.00016099394771323393, + "loss": 0.0553, + "step": 37510 + }, + { + "epoch": 0.5852623697510451, + "grad_norm": 0.11538795381784439, + "learning_rate": 0.00016098354859507913, + "loss": 0.0889, + "step": 37520 + }, + { + "epoch": 0.5854183565233668, + "grad_norm": 1.0372689962387085, + "learning_rate": 0.00016097314947692437, + "loss": 0.3909, + "step": 37530 + }, + { + "epoch": 0.5855743432956885, + "grad_norm": 1.1946550607681274, + "learning_rate": 0.00016096275035876957, + "loss": 0.2167, + "step": 37540 + }, + { + "epoch": 0.5857303300680102, + "grad_norm": 1.3474591970443726, + "learning_rate": 0.0001609523512406148, + "loss": 0.1561, + "step": 37550 + }, + { + "epoch": 0.585886316840332, + "grad_norm": 0.008936616592109203, + "learning_rate": 0.00016094195212246, + "loss": 0.3991, + "step": 37560 + }, + { + "epoch": 0.5860423036126536, + "grad_norm": 1.8822320699691772, + "learning_rate": 0.00016093155300430525, + "loss": 0.2438, + "step": 37570 + }, + { + "epoch": 0.5861982903849754, + "grad_norm": 1.1707524061203003, + "learning_rate": 0.00016092115388615044, + "loss": 0.308, + "step": 37580 + }, + { + "epoch": 0.586354277157297, + "grad_norm": 1.679641842842102, + "learning_rate": 0.0001609107547679957, + "loss": 0.159, + "step": 37590 + }, + { + "epoch": 0.5865102639296188, + "grad_norm": 1.6885284185409546, + "learning_rate": 0.00016090035564984088, + "loss": 0.157, + "step": 37600 + }, + { + "epoch": 0.5866662507019405, + "grad_norm": 0.3876996636390686, + "learning_rate": 0.00016088995653168613, + "loss": 0.4067, + "step": 37610 + }, + { + "epoch": 0.5868222374742622, + "grad_norm": 1.387480616569519, + "learning_rate": 0.00016087955741353132, + "loss": 0.1898, + "step": 37620 + }, + { + "epoch": 0.5869782242465839, + "grad_norm": 0.03653861582279205, + "learning_rate": 0.00016086915829537657, + "loss": 0.1652, + "step": 37630 + }, + { + "epoch": 0.5871342110189056, + "grad_norm": 0.6258412003517151, + "learning_rate": 0.00016085875917722176, + "loss": 0.2558, + "step": 37640 + }, + { + "epoch": 0.5872901977912273, + "grad_norm": 0.10519164055585861, + "learning_rate": 0.000160848360059067, + "loss": 0.243, + "step": 37650 + }, + { + "epoch": 0.587446184563549, + "grad_norm": 0.7370131015777588, + "learning_rate": 0.0001608379609409122, + "loss": 0.2343, + "step": 37660 + }, + { + "epoch": 0.5876021713358707, + "grad_norm": 0.6884517073631287, + "learning_rate": 0.00016082756182275745, + "loss": 0.1847, + "step": 37670 + }, + { + "epoch": 0.5877581581081924, + "grad_norm": 0.6326120495796204, + "learning_rate": 0.00016081716270460264, + "loss": 0.2258, + "step": 37680 + }, + { + "epoch": 0.5879141448805141, + "grad_norm": 7.697889804840088, + "learning_rate": 0.0001608067635864479, + "loss": 0.2068, + "step": 37690 + }, + { + "epoch": 0.5880701316528358, + "grad_norm": 1.8753420114517212, + "learning_rate": 0.00016079636446829308, + "loss": 0.3087, + "step": 37700 + }, + { + "epoch": 0.5882261184251576, + "grad_norm": 0.8591625094413757, + "learning_rate": 0.00016078596535013833, + "loss": 0.1082, + "step": 37710 + }, + { + "epoch": 0.5883821051974792, + "grad_norm": 0.6124423146247864, + "learning_rate": 0.00016077556623198352, + "loss": 0.1536, + "step": 37720 + }, + { + "epoch": 0.588538091969801, + "grad_norm": 0.1400165557861328, + "learning_rate": 0.00016076516711382877, + "loss": 0.145, + "step": 37730 + }, + { + "epoch": 0.5886940787421227, + "grad_norm": 1.9879004955291748, + "learning_rate": 0.00016075476799567396, + "loss": 0.4986, + "step": 37740 + }, + { + "epoch": 0.5888500655144444, + "grad_norm": 0.08687327802181244, + "learning_rate": 0.0001607443688775192, + "loss": 0.3365, + "step": 37750 + }, + { + "epoch": 0.5890060522867661, + "grad_norm": 0.3442278206348419, + "learning_rate": 0.0001607339697593644, + "loss": 0.2274, + "step": 37760 + }, + { + "epoch": 0.5891620390590878, + "grad_norm": 0.06124915927648544, + "learning_rate": 0.00016072357064120965, + "loss": 0.1915, + "step": 37770 + }, + { + "epoch": 0.5893180258314095, + "grad_norm": 1.0996816158294678, + "learning_rate": 0.00016071317152305484, + "loss": 0.1573, + "step": 37780 + }, + { + "epoch": 0.5894740126037312, + "grad_norm": 0.14738436043262482, + "learning_rate": 0.00016070277240490008, + "loss": 0.2524, + "step": 37790 + }, + { + "epoch": 0.5896299993760529, + "grad_norm": 2.3905467987060547, + "learning_rate": 0.00016069237328674528, + "loss": 0.1838, + "step": 37800 + }, + { + "epoch": 0.5897859861483746, + "grad_norm": 2.0800178050994873, + "learning_rate": 0.00016068197416859052, + "loss": 0.3749, + "step": 37810 + }, + { + "epoch": 0.5899419729206963, + "grad_norm": 3.254204511642456, + "learning_rate": 0.00016067157505043574, + "loss": 0.3777, + "step": 37820 + }, + { + "epoch": 0.590097959693018, + "grad_norm": 2.8447892665863037, + "learning_rate": 0.00016066117593228096, + "loss": 0.5338, + "step": 37830 + }, + { + "epoch": 0.5902539464653397, + "grad_norm": 0.624733030796051, + "learning_rate": 0.00016065077681412618, + "loss": 0.2331, + "step": 37840 + }, + { + "epoch": 0.5904099332376614, + "grad_norm": 0.6863507628440857, + "learning_rate": 0.0001606403776959714, + "loss": 0.1747, + "step": 37850 + }, + { + "epoch": 0.5905659200099832, + "grad_norm": 0.6785943508148193, + "learning_rate": 0.00016062997857781662, + "loss": 0.2064, + "step": 37860 + }, + { + "epoch": 0.5907219067823049, + "grad_norm": 0.5739938616752625, + "learning_rate": 0.00016061957945966181, + "loss": 0.2368, + "step": 37870 + }, + { + "epoch": 0.5908778935546266, + "grad_norm": 0.0616387277841568, + "learning_rate": 0.00016060918034150706, + "loss": 0.2224, + "step": 37880 + }, + { + "epoch": 0.5910338803269483, + "grad_norm": 2.745957136154175, + "learning_rate": 0.00016059878122335225, + "loss": 0.3161, + "step": 37890 + }, + { + "epoch": 0.59118986709927, + "grad_norm": 2.52472186088562, + "learning_rate": 0.0001605883821051975, + "loss": 0.1713, + "step": 37900 + }, + { + "epoch": 0.5913458538715917, + "grad_norm": 1.5691877603530884, + "learning_rate": 0.0001605779829870427, + "loss": 0.1402, + "step": 37910 + }, + { + "epoch": 0.5915018406439134, + "grad_norm": 1.848261833190918, + "learning_rate": 0.00016056758386888794, + "loss": 0.2305, + "step": 37920 + }, + { + "epoch": 0.5916578274162351, + "grad_norm": 0.36622482538223267, + "learning_rate": 0.00016055718475073313, + "loss": 0.1834, + "step": 37930 + }, + { + "epoch": 0.5918138141885568, + "grad_norm": 0.21201446652412415, + "learning_rate": 0.00016054678563257838, + "loss": 0.4682, + "step": 37940 + }, + { + "epoch": 0.5919698009608785, + "grad_norm": 0.38921502232551575, + "learning_rate": 0.00016053638651442357, + "loss": 0.2892, + "step": 37950 + }, + { + "epoch": 0.5921257877332002, + "grad_norm": 0.20777840912342072, + "learning_rate": 0.00016052598739626882, + "loss": 0.104, + "step": 37960 + }, + { + "epoch": 0.5922817745055219, + "grad_norm": 0.7090150713920593, + "learning_rate": 0.000160515588278114, + "loss": 0.2592, + "step": 37970 + }, + { + "epoch": 0.5924377612778436, + "grad_norm": 1.9939709901809692, + "learning_rate": 0.00016050518915995926, + "loss": 0.247, + "step": 37980 + }, + { + "epoch": 0.5925937480501653, + "grad_norm": 0.8696603178977966, + "learning_rate": 0.00016049479004180445, + "loss": 0.301, + "step": 37990 + }, + { + "epoch": 0.592749734822487, + "grad_norm": 0.4119095206260681, + "learning_rate": 0.0001604843909236497, + "loss": 0.2011, + "step": 38000 + }, + { + "epoch": 0.5929057215948088, + "grad_norm": 0.060975607484579086, + "learning_rate": 0.0001604739918054949, + "loss": 0.2286, + "step": 38010 + }, + { + "epoch": 0.5930617083671305, + "grad_norm": 0.7039875388145447, + "learning_rate": 0.00016046359268734014, + "loss": 0.0999, + "step": 38020 + }, + { + "epoch": 0.5932176951394522, + "grad_norm": 0.28402045369148254, + "learning_rate": 0.00016045319356918533, + "loss": 0.1835, + "step": 38030 + }, + { + "epoch": 0.5933736819117739, + "grad_norm": 2.217988967895508, + "learning_rate": 0.00016044279445103058, + "loss": 0.0785, + "step": 38040 + }, + { + "epoch": 0.5935296686840956, + "grad_norm": 1.4856069087982178, + "learning_rate": 0.00016043239533287577, + "loss": 0.2528, + "step": 38050 + }, + { + "epoch": 0.5936856554564173, + "grad_norm": 0.9875519871711731, + "learning_rate": 0.00016042199621472101, + "loss": 0.4445, + "step": 38060 + }, + { + "epoch": 0.593841642228739, + "grad_norm": 2.773216724395752, + "learning_rate": 0.0001604115970965662, + "loss": 0.5643, + "step": 38070 + }, + { + "epoch": 0.5939976290010607, + "grad_norm": 1.7584549188613892, + "learning_rate": 0.00016040119797841145, + "loss": 0.1446, + "step": 38080 + }, + { + "epoch": 0.5941536157733824, + "grad_norm": 1.2456908226013184, + "learning_rate": 0.00016039079886025665, + "loss": 0.1896, + "step": 38090 + }, + { + "epoch": 0.5943096025457041, + "grad_norm": 0.021098516881465912, + "learning_rate": 0.0001603803997421019, + "loss": 0.0876, + "step": 38100 + }, + { + "epoch": 0.5944655893180258, + "grad_norm": 1.8131749629974365, + "learning_rate": 0.00016037000062394709, + "loss": 0.1639, + "step": 38110 + }, + { + "epoch": 0.5946215760903475, + "grad_norm": 0.41819462180137634, + "learning_rate": 0.00016035960150579233, + "loss": 0.1311, + "step": 38120 + }, + { + "epoch": 0.5947775628626693, + "grad_norm": 0.6691219806671143, + "learning_rate": 0.00016034920238763753, + "loss": 0.1625, + "step": 38130 + }, + { + "epoch": 0.5949335496349909, + "grad_norm": 0.07854852080345154, + "learning_rate": 0.00016033880326948277, + "loss": 0.3177, + "step": 38140 + }, + { + "epoch": 0.5950895364073127, + "grad_norm": 0.7819689512252808, + "learning_rate": 0.00016032840415132796, + "loss": 0.4225, + "step": 38150 + }, + { + "epoch": 0.5952455231796344, + "grad_norm": 5.3962836265563965, + "learning_rate": 0.0001603180050331732, + "loss": 0.123, + "step": 38160 + }, + { + "epoch": 0.5954015099519561, + "grad_norm": 0.7460303902626038, + "learning_rate": 0.0001603076059150184, + "loss": 0.2182, + "step": 38170 + }, + { + "epoch": 0.5955574967242778, + "grad_norm": 0.26433685421943665, + "learning_rate": 0.00016029720679686365, + "loss": 0.1894, + "step": 38180 + }, + { + "epoch": 0.5957134834965995, + "grad_norm": 0.8701320290565491, + "learning_rate": 0.00016028680767870884, + "loss": 0.2137, + "step": 38190 + }, + { + "epoch": 0.5958694702689212, + "grad_norm": 3.9774930477142334, + "learning_rate": 0.0001602764085605541, + "loss": 0.4637, + "step": 38200 + }, + { + "epoch": 0.5960254570412429, + "grad_norm": 0.5801145434379578, + "learning_rate": 0.00016026600944239928, + "loss": 0.3036, + "step": 38210 + }, + { + "epoch": 0.5961814438135646, + "grad_norm": 1.1890723705291748, + "learning_rate": 0.00016025561032424453, + "loss": 0.0988, + "step": 38220 + }, + { + "epoch": 0.5963374305858863, + "grad_norm": 0.18521073460578918, + "learning_rate": 0.00016024521120608972, + "loss": 0.1529, + "step": 38230 + }, + { + "epoch": 0.596493417358208, + "grad_norm": 2.2602827548980713, + "learning_rate": 0.00016023481208793497, + "loss": 0.2815, + "step": 38240 + }, + { + "epoch": 0.5966494041305297, + "grad_norm": 0.4060989022254944, + "learning_rate": 0.00016022441296978016, + "loss": 0.3466, + "step": 38250 + }, + { + "epoch": 0.5968053909028515, + "grad_norm": 1.4743244647979736, + "learning_rate": 0.0001602140138516254, + "loss": 0.3561, + "step": 38260 + }, + { + "epoch": 0.5969613776751731, + "grad_norm": 5.080365180969238, + "learning_rate": 0.0001602036147334706, + "loss": 0.2914, + "step": 38270 + }, + { + "epoch": 0.5971173644474949, + "grad_norm": 0.16149812936782837, + "learning_rate": 0.00016019321561531585, + "loss": 0.1871, + "step": 38280 + }, + { + "epoch": 0.5972733512198165, + "grad_norm": 1.7535364627838135, + "learning_rate": 0.00016018281649716104, + "loss": 0.4905, + "step": 38290 + }, + { + "epoch": 0.5974293379921383, + "grad_norm": 0.4103987216949463, + "learning_rate": 0.00016017241737900629, + "loss": 0.3813, + "step": 38300 + }, + { + "epoch": 0.59758532476446, + "grad_norm": 0.5315369367599487, + "learning_rate": 0.00016016201826085148, + "loss": 0.3897, + "step": 38310 + }, + { + "epoch": 0.5977413115367817, + "grad_norm": 3.4553463459014893, + "learning_rate": 0.0001601516191426967, + "loss": 0.2047, + "step": 38320 + }, + { + "epoch": 0.5978972983091034, + "grad_norm": 2.225468397140503, + "learning_rate": 0.00016014122002454192, + "loss": 0.4028, + "step": 38330 + }, + { + "epoch": 0.5980532850814251, + "grad_norm": 2.000885486602783, + "learning_rate": 0.00016013082090638714, + "loss": 0.391, + "step": 38340 + }, + { + "epoch": 0.5982092718537468, + "grad_norm": 0.7886672616004944, + "learning_rate": 0.00016012042178823236, + "loss": 0.2994, + "step": 38350 + }, + { + "epoch": 0.5983652586260685, + "grad_norm": 0.5279917120933533, + "learning_rate": 0.00016011002267007758, + "loss": 0.1855, + "step": 38360 + }, + { + "epoch": 0.5985212453983902, + "grad_norm": 0.9789942502975464, + "learning_rate": 0.0001600996235519228, + "loss": 0.2122, + "step": 38370 + }, + { + "epoch": 0.5986772321707119, + "grad_norm": 0.6289750933647156, + "learning_rate": 0.00016008922443376802, + "loss": 0.3737, + "step": 38380 + }, + { + "epoch": 0.5988332189430337, + "grad_norm": 1.3074733018875122, + "learning_rate": 0.00016007882531561324, + "loss": 0.2382, + "step": 38390 + }, + { + "epoch": 0.5989892057153553, + "grad_norm": 1.745284914970398, + "learning_rate": 0.00016006842619745846, + "loss": 0.3167, + "step": 38400 + }, + { + "epoch": 0.5991451924876771, + "grad_norm": 0.1694452464580536, + "learning_rate": 0.00016005802707930368, + "loss": 0.1632, + "step": 38410 + }, + { + "epoch": 0.5993011792599987, + "grad_norm": 1.55560302734375, + "learning_rate": 0.0001600476279611489, + "loss": 0.2371, + "step": 38420 + }, + { + "epoch": 0.5994571660323205, + "grad_norm": 1.044997215270996, + "learning_rate": 0.00016003722884299411, + "loss": 0.0896, + "step": 38430 + }, + { + "epoch": 0.5996131528046421, + "grad_norm": 0.25250110030174255, + "learning_rate": 0.00016002682972483933, + "loss": 0.3096, + "step": 38440 + }, + { + "epoch": 0.5997691395769639, + "grad_norm": 0.1631859689950943, + "learning_rate": 0.00016001643060668455, + "loss": 0.2004, + "step": 38450 + }, + { + "epoch": 0.5999251263492856, + "grad_norm": 0.34967318177223206, + "learning_rate": 0.00016000603148852977, + "loss": 0.1476, + "step": 38460 + }, + { + "epoch": 0.6000811131216073, + "grad_norm": 1.8604000806808472, + "learning_rate": 0.000159995632370375, + "loss": 0.1884, + "step": 38470 + }, + { + "epoch": 0.600237099893929, + "grad_norm": 1.626848578453064, + "learning_rate": 0.0001599852332522202, + "loss": 0.5452, + "step": 38480 + }, + { + "epoch": 0.6003930866662507, + "grad_norm": 0.41935989260673523, + "learning_rate": 0.00015997483413406543, + "loss": 0.3047, + "step": 38490 + }, + { + "epoch": 0.6005490734385724, + "grad_norm": 1.355086088180542, + "learning_rate": 0.00015996443501591065, + "loss": 0.1246, + "step": 38500 + }, + { + "epoch": 0.6007050602108941, + "grad_norm": 1.5382829904556274, + "learning_rate": 0.00015995403589775587, + "loss": 0.2324, + "step": 38510 + }, + { + "epoch": 0.6008610469832159, + "grad_norm": 1.0250797271728516, + "learning_rate": 0.0001599436367796011, + "loss": 0.181, + "step": 38520 + }, + { + "epoch": 0.6010170337555375, + "grad_norm": 0.07823364436626434, + "learning_rate": 0.0001599332376614463, + "loss": 0.1492, + "step": 38530 + }, + { + "epoch": 0.6011730205278593, + "grad_norm": 0.07573071867227554, + "learning_rate": 0.00015992283854329153, + "loss": 0.1256, + "step": 38540 + }, + { + "epoch": 0.6013290073001809, + "grad_norm": 0.5608569383621216, + "learning_rate": 0.00015991243942513675, + "loss": 0.378, + "step": 38550 + }, + { + "epoch": 0.6014849940725027, + "grad_norm": 0.7001075744628906, + "learning_rate": 0.00015990204030698197, + "loss": 0.2518, + "step": 38560 + }, + { + "epoch": 0.6016409808448243, + "grad_norm": 4.868145942687988, + "learning_rate": 0.0001598916411888272, + "loss": 0.1857, + "step": 38570 + }, + { + "epoch": 0.6017969676171461, + "grad_norm": 0.05653705075383186, + "learning_rate": 0.0001598812420706724, + "loss": 0.1918, + "step": 38580 + }, + { + "epoch": 0.6019529543894677, + "grad_norm": 0.8556832671165466, + "learning_rate": 0.00015987084295251763, + "loss": 0.1926, + "step": 38590 + }, + { + "epoch": 0.6021089411617895, + "grad_norm": 1.1988335847854614, + "learning_rate": 0.00015986044383436285, + "loss": 0.3884, + "step": 38600 + }, + { + "epoch": 0.6022649279341112, + "grad_norm": 1.5978832244873047, + "learning_rate": 0.00015985004471620807, + "loss": 0.1986, + "step": 38610 + }, + { + "epoch": 0.6024209147064329, + "grad_norm": 1.084174633026123, + "learning_rate": 0.0001598396455980533, + "loss": 0.4602, + "step": 38620 + }, + { + "epoch": 0.6025769014787546, + "grad_norm": 1.5116629600524902, + "learning_rate": 0.0001598292464798985, + "loss": 0.292, + "step": 38630 + }, + { + "epoch": 0.6027328882510763, + "grad_norm": 0.23945322632789612, + "learning_rate": 0.00015981884736174373, + "loss": 0.1904, + "step": 38640 + }, + { + "epoch": 0.602888875023398, + "grad_norm": 0.348156601190567, + "learning_rate": 0.00015980844824358895, + "loss": 0.1442, + "step": 38650 + }, + { + "epoch": 0.6030448617957197, + "grad_norm": 1.876736044883728, + "learning_rate": 0.00015979804912543417, + "loss": 0.248, + "step": 38660 + }, + { + "epoch": 0.6032008485680415, + "grad_norm": 0.2564373016357422, + "learning_rate": 0.00015978765000727939, + "loss": 0.2429, + "step": 38670 + }, + { + "epoch": 0.6033568353403631, + "grad_norm": 1.824837327003479, + "learning_rate": 0.0001597772508891246, + "loss": 0.2593, + "step": 38680 + }, + { + "epoch": 0.6035128221126849, + "grad_norm": 1.6967028379440308, + "learning_rate": 0.00015976685177096985, + "loss": 0.3918, + "step": 38690 + }, + { + "epoch": 0.6036688088850065, + "grad_norm": 1.6629718542099, + "learning_rate": 0.00015975645265281504, + "loss": 0.2529, + "step": 38700 + }, + { + "epoch": 0.6038247956573283, + "grad_norm": 0.19047370553016663, + "learning_rate": 0.0001597460535346603, + "loss": 0.4545, + "step": 38710 + }, + { + "epoch": 0.6039807824296499, + "grad_norm": 0.6910511255264282, + "learning_rate": 0.00015973565441650548, + "loss": 0.2226, + "step": 38720 + }, + { + "epoch": 0.6041367692019717, + "grad_norm": 1.9387328624725342, + "learning_rate": 0.00015972525529835073, + "loss": 0.5561, + "step": 38730 + }, + { + "epoch": 0.6042927559742933, + "grad_norm": 1.4491708278656006, + "learning_rate": 0.00015971485618019592, + "loss": 0.2987, + "step": 38740 + }, + { + "epoch": 0.6044487427466151, + "grad_norm": 0.32534459233283997, + "learning_rate": 0.00015970445706204117, + "loss": 0.1683, + "step": 38750 + }, + { + "epoch": 0.6046047295189368, + "grad_norm": 0.17335011065006256, + "learning_rate": 0.00015969405794388636, + "loss": 0.1554, + "step": 38760 + }, + { + "epoch": 0.6047607162912585, + "grad_norm": 0.39481663703918457, + "learning_rate": 0.00015968365882573158, + "loss": 0.186, + "step": 38770 + }, + { + "epoch": 0.6049167030635803, + "grad_norm": 0.10422719269990921, + "learning_rate": 0.0001596732597075768, + "loss": 0.1599, + "step": 38780 + }, + { + "epoch": 0.6050726898359019, + "grad_norm": 2.288064479827881, + "learning_rate": 0.00015966286058942202, + "loss": 0.1944, + "step": 38790 + }, + { + "epoch": 0.6052286766082237, + "grad_norm": 1.0062401294708252, + "learning_rate": 0.00015965246147126724, + "loss": 0.176, + "step": 38800 + }, + { + "epoch": 0.6053846633805453, + "grad_norm": 0.21470843255519867, + "learning_rate": 0.00015964206235311246, + "loss": 0.183, + "step": 38810 + }, + { + "epoch": 0.6055406501528671, + "grad_norm": 2.1730270385742188, + "learning_rate": 0.00015963166323495768, + "loss": 0.1788, + "step": 38820 + }, + { + "epoch": 0.6056966369251887, + "grad_norm": 0.6188749074935913, + "learning_rate": 0.0001596212641168029, + "loss": 0.2746, + "step": 38830 + }, + { + "epoch": 0.6058526236975105, + "grad_norm": 1.283058524131775, + "learning_rate": 0.00015961086499864812, + "loss": 0.1066, + "step": 38840 + }, + { + "epoch": 0.6060086104698321, + "grad_norm": 1.1469789743423462, + "learning_rate": 0.00015960046588049334, + "loss": 0.0692, + "step": 38850 + }, + { + "epoch": 0.6061645972421539, + "grad_norm": 1.9450418949127197, + "learning_rate": 0.00015959006676233856, + "loss": 0.2154, + "step": 38860 + }, + { + "epoch": 0.6063205840144755, + "grad_norm": 7.193401336669922, + "learning_rate": 0.00015957966764418378, + "loss": 0.218, + "step": 38870 + }, + { + "epoch": 0.6064765707867973, + "grad_norm": 3.5315003395080566, + "learning_rate": 0.000159569268526029, + "loss": 0.312, + "step": 38880 + }, + { + "epoch": 0.6066325575591189, + "grad_norm": 1.2873650789260864, + "learning_rate": 0.00015955886940787422, + "loss": 0.2587, + "step": 38890 + }, + { + "epoch": 0.6067885443314407, + "grad_norm": 1.1193941831588745, + "learning_rate": 0.00015954847028971944, + "loss": 0.2358, + "step": 38900 + }, + { + "epoch": 0.6069445311037625, + "grad_norm": 1.0471165180206299, + "learning_rate": 0.00015953807117156466, + "loss": 0.2639, + "step": 38910 + }, + { + "epoch": 0.6071005178760841, + "grad_norm": 1.7760034799575806, + "learning_rate": 0.00015952767205340988, + "loss": 0.248, + "step": 38920 + }, + { + "epoch": 0.6072565046484059, + "grad_norm": 0.050742994993925095, + "learning_rate": 0.0001595172729352551, + "loss": 0.1287, + "step": 38930 + }, + { + "epoch": 0.6074124914207275, + "grad_norm": 0.03016967698931694, + "learning_rate": 0.00015950687381710032, + "loss": 0.1843, + "step": 38940 + }, + { + "epoch": 0.6075684781930493, + "grad_norm": 0.6029847860336304, + "learning_rate": 0.00015949647469894554, + "loss": 0.2294, + "step": 38950 + }, + { + "epoch": 0.6077244649653709, + "grad_norm": 0.2955610454082489, + "learning_rate": 0.00015948607558079076, + "loss": 0.3336, + "step": 38960 + }, + { + "epoch": 0.6078804517376927, + "grad_norm": 1.9104161262512207, + "learning_rate": 0.00015947567646263597, + "loss": 0.189, + "step": 38970 + }, + { + "epoch": 0.6080364385100143, + "grad_norm": 0.09197711199522018, + "learning_rate": 0.0001594652773444812, + "loss": 0.1612, + "step": 38980 + }, + { + "epoch": 0.6081924252823361, + "grad_norm": 1.4233628511428833, + "learning_rate": 0.00015945487822632641, + "loss": 0.1521, + "step": 38990 + }, + { + "epoch": 0.6083484120546577, + "grad_norm": 9.787493705749512, + "learning_rate": 0.00015944447910817163, + "loss": 0.135, + "step": 39000 + }, + { + "epoch": 0.6085043988269795, + "grad_norm": 1.3848742246627808, + "learning_rate": 0.00015943407999001685, + "loss": 0.2268, + "step": 39010 + }, + { + "epoch": 0.6086603855993011, + "grad_norm": 0.1524626612663269, + "learning_rate": 0.00015942368087186207, + "loss": 0.3154, + "step": 39020 + }, + { + "epoch": 0.6088163723716229, + "grad_norm": 2.1885435581207275, + "learning_rate": 0.0001594132817537073, + "loss": 0.3174, + "step": 39030 + }, + { + "epoch": 0.6089723591439445, + "grad_norm": 1.4719898700714111, + "learning_rate": 0.0001594028826355525, + "loss": 0.2959, + "step": 39040 + }, + { + "epoch": 0.6091283459162663, + "grad_norm": 0.7844040989875793, + "learning_rate": 0.00015939248351739773, + "loss": 0.3021, + "step": 39050 + }, + { + "epoch": 0.6092843326885881, + "grad_norm": 0.17660890519618988, + "learning_rate": 0.00015938208439924295, + "loss": 0.3608, + "step": 39060 + }, + { + "epoch": 0.6094403194609097, + "grad_norm": 0.24621741473674774, + "learning_rate": 0.00015937168528108817, + "loss": 0.1461, + "step": 39070 + }, + { + "epoch": 0.6095963062332315, + "grad_norm": 0.7903050184249878, + "learning_rate": 0.0001593612861629334, + "loss": 0.1195, + "step": 39080 + }, + { + "epoch": 0.6097522930055531, + "grad_norm": 0.04043950140476227, + "learning_rate": 0.0001593508870447786, + "loss": 0.1938, + "step": 39090 + }, + { + "epoch": 0.6099082797778749, + "grad_norm": 0.7602111101150513, + "learning_rate": 0.00015934048792662383, + "loss": 0.1732, + "step": 39100 + }, + { + "epoch": 0.6100642665501965, + "grad_norm": 0.7902756333351135, + "learning_rate": 0.00015933008880846905, + "loss": 0.2351, + "step": 39110 + }, + { + "epoch": 0.6102202533225183, + "grad_norm": 1.7035608291625977, + "learning_rate": 0.00015931968969031427, + "loss": 0.2409, + "step": 39120 + }, + { + "epoch": 0.6103762400948399, + "grad_norm": 0.7998217940330505, + "learning_rate": 0.0001593092905721595, + "loss": 0.3862, + "step": 39130 + }, + { + "epoch": 0.6105322268671617, + "grad_norm": 0.3530103862285614, + "learning_rate": 0.0001592988914540047, + "loss": 0.2176, + "step": 39140 + }, + { + "epoch": 0.6106882136394833, + "grad_norm": 0.9218409657478333, + "learning_rate": 0.00015928849233584993, + "loss": 0.1362, + "step": 39150 + }, + { + "epoch": 0.6108442004118051, + "grad_norm": 0.643829345703125, + "learning_rate": 0.00015927809321769515, + "loss": 0.2309, + "step": 39160 + }, + { + "epoch": 0.6110001871841267, + "grad_norm": 0.1810176521539688, + "learning_rate": 0.00015926769409954037, + "loss": 0.2461, + "step": 39170 + }, + { + "epoch": 0.6111561739564485, + "grad_norm": 1.3435649871826172, + "learning_rate": 0.0001592572949813856, + "loss": 0.2937, + "step": 39180 + }, + { + "epoch": 0.6113121607287701, + "grad_norm": 1.1048580408096313, + "learning_rate": 0.0001592468958632308, + "loss": 0.2983, + "step": 39190 + }, + { + "epoch": 0.6114681475010919, + "grad_norm": 0.09463644027709961, + "learning_rate": 0.00015923649674507603, + "loss": 0.2318, + "step": 39200 + }, + { + "epoch": 0.6116241342734137, + "grad_norm": 0.9329594969749451, + "learning_rate": 0.00015922609762692125, + "loss": 0.3095, + "step": 39210 + }, + { + "epoch": 0.6117801210457353, + "grad_norm": 1.8576663732528687, + "learning_rate": 0.00015921569850876647, + "loss": 0.2336, + "step": 39220 + }, + { + "epoch": 0.6119361078180571, + "grad_norm": 1.6159908771514893, + "learning_rate": 0.00015920529939061169, + "loss": 0.1745, + "step": 39230 + }, + { + "epoch": 0.6120920945903787, + "grad_norm": 1.289712905883789, + "learning_rate": 0.0001591949002724569, + "loss": 0.2031, + "step": 39240 + }, + { + "epoch": 0.6122480813627005, + "grad_norm": 0.4183761179447174, + "learning_rate": 0.00015918450115430212, + "loss": 0.2911, + "step": 39250 + }, + { + "epoch": 0.6124040681350221, + "grad_norm": 2.125042200088501, + "learning_rate": 0.00015917410203614734, + "loss": 0.1086, + "step": 39260 + }, + { + "epoch": 0.6125600549073439, + "grad_norm": 3.527683973312378, + "learning_rate": 0.00015916370291799256, + "loss": 0.3356, + "step": 39270 + }, + { + "epoch": 0.6127160416796655, + "grad_norm": 1.1525741815567017, + "learning_rate": 0.00015915330379983778, + "loss": 0.222, + "step": 39280 + }, + { + "epoch": 0.6128720284519873, + "grad_norm": 0.5210650563240051, + "learning_rate": 0.000159142904681683, + "loss": 0.1563, + "step": 39290 + }, + { + "epoch": 0.6130280152243089, + "grad_norm": 0.2881133258342743, + "learning_rate": 0.00015913250556352822, + "loss": 0.346, + "step": 39300 + }, + { + "epoch": 0.6131840019966307, + "grad_norm": 1.7034498453140259, + "learning_rate": 0.00015912210644537344, + "loss": 0.1822, + "step": 39310 + }, + { + "epoch": 0.6133399887689523, + "grad_norm": 2.6305196285247803, + "learning_rate": 0.00015911170732721866, + "loss": 0.3373, + "step": 39320 + }, + { + "epoch": 0.6134959755412741, + "grad_norm": 0.4541589915752411, + "learning_rate": 0.00015910130820906388, + "loss": 0.2146, + "step": 39330 + }, + { + "epoch": 0.6136519623135958, + "grad_norm": 0.03301222622394562, + "learning_rate": 0.0001590909090909091, + "loss": 0.1787, + "step": 39340 + }, + { + "epoch": 0.6138079490859175, + "grad_norm": 0.7106437087059021, + "learning_rate": 0.00015908050997275432, + "loss": 0.2315, + "step": 39350 + }, + { + "epoch": 0.6139639358582393, + "grad_norm": 0.381320059299469, + "learning_rate": 0.00015907011085459954, + "loss": 0.1917, + "step": 39360 + }, + { + "epoch": 0.6141199226305609, + "grad_norm": 0.5418739914894104, + "learning_rate": 0.00015905971173644476, + "loss": 0.1571, + "step": 39370 + }, + { + "epoch": 0.6142759094028827, + "grad_norm": 0.3135250210762024, + "learning_rate": 0.00015904931261828998, + "loss": 0.0858, + "step": 39380 + }, + { + "epoch": 0.6144318961752043, + "grad_norm": 1.6296218633651733, + "learning_rate": 0.0001590389135001352, + "loss": 0.2297, + "step": 39390 + }, + { + "epoch": 0.6145878829475261, + "grad_norm": 0.5768391489982605, + "learning_rate": 0.00015902851438198042, + "loss": 0.1699, + "step": 39400 + }, + { + "epoch": 0.6147438697198477, + "grad_norm": 25.794370651245117, + "learning_rate": 0.00015901811526382564, + "loss": 0.1903, + "step": 39410 + }, + { + "epoch": 0.6148998564921695, + "grad_norm": 0.46698322892189026, + "learning_rate": 0.00015900771614567086, + "loss": 0.1415, + "step": 39420 + }, + { + "epoch": 0.6150558432644911, + "grad_norm": 1.3042336702346802, + "learning_rate": 0.00015899731702751608, + "loss": 0.1604, + "step": 39430 + }, + { + "epoch": 0.6152118300368129, + "grad_norm": 1.2311556339263916, + "learning_rate": 0.0001589869179093613, + "loss": 0.2555, + "step": 39440 + }, + { + "epoch": 0.6153678168091345, + "grad_norm": 0.6290935277938843, + "learning_rate": 0.00015897651879120652, + "loss": 0.1723, + "step": 39450 + }, + { + "epoch": 0.6155238035814563, + "grad_norm": 1.4786419868469238, + "learning_rate": 0.00015896611967305174, + "loss": 0.3717, + "step": 39460 + }, + { + "epoch": 0.615679790353778, + "grad_norm": 0.7589457631111145, + "learning_rate": 0.00015895572055489696, + "loss": 0.1912, + "step": 39470 + }, + { + "epoch": 0.6158357771260997, + "grad_norm": 0.10508652776479721, + "learning_rate": 0.00015894532143674218, + "loss": 0.1353, + "step": 39480 + }, + { + "epoch": 0.6159917638984214, + "grad_norm": 0.07394446432590485, + "learning_rate": 0.0001589349223185874, + "loss": 0.1195, + "step": 39490 + }, + { + "epoch": 0.6161477506707431, + "grad_norm": 0.11420007050037384, + "learning_rate": 0.00015892452320043262, + "loss": 0.0976, + "step": 39500 + }, + { + "epoch": 0.6163037374430649, + "grad_norm": 1.2442359924316406, + "learning_rate": 0.00015891412408227784, + "loss": 0.1084, + "step": 39510 + }, + { + "epoch": 0.6164597242153865, + "grad_norm": 0.4080588221549988, + "learning_rate": 0.00015890372496412306, + "loss": 0.2191, + "step": 39520 + }, + { + "epoch": 0.6166157109877083, + "grad_norm": 0.25214430689811707, + "learning_rate": 0.00015889332584596827, + "loss": 0.5548, + "step": 39530 + }, + { + "epoch": 0.6167716977600299, + "grad_norm": 0.41108590364456177, + "learning_rate": 0.0001588829267278135, + "loss": 0.4213, + "step": 39540 + }, + { + "epoch": 0.6169276845323517, + "grad_norm": 3.3010945320129395, + "learning_rate": 0.00015887252760965871, + "loss": 0.3946, + "step": 39550 + }, + { + "epoch": 0.6170836713046733, + "grad_norm": 0.25975245237350464, + "learning_rate": 0.00015886212849150393, + "loss": 0.3045, + "step": 39560 + }, + { + "epoch": 0.6172396580769951, + "grad_norm": 1.255585789680481, + "learning_rate": 0.00015885172937334915, + "loss": 0.2249, + "step": 39570 + }, + { + "epoch": 0.6173956448493167, + "grad_norm": 1.9135524034500122, + "learning_rate": 0.00015884133025519437, + "loss": 0.2401, + "step": 39580 + }, + { + "epoch": 0.6175516316216385, + "grad_norm": 2.1866586208343506, + "learning_rate": 0.0001588309311370396, + "loss": 0.1726, + "step": 39590 + }, + { + "epoch": 0.6177076183939602, + "grad_norm": 2.487704277038574, + "learning_rate": 0.0001588205320188848, + "loss": 0.2728, + "step": 39600 + }, + { + "epoch": 0.6178636051662819, + "grad_norm": 0.7800028920173645, + "learning_rate": 0.00015881013290073003, + "loss": 0.1522, + "step": 39610 + }, + { + "epoch": 0.6180195919386036, + "grad_norm": 3.468012571334839, + "learning_rate": 0.00015879973378257525, + "loss": 0.1984, + "step": 39620 + }, + { + "epoch": 0.6181755787109253, + "grad_norm": 1.5809861421585083, + "learning_rate": 0.00015878933466442047, + "loss": 0.2632, + "step": 39630 + }, + { + "epoch": 0.618331565483247, + "grad_norm": 0.5371276140213013, + "learning_rate": 0.0001587789355462657, + "loss": 0.1915, + "step": 39640 + }, + { + "epoch": 0.6184875522555687, + "grad_norm": 4.711477279663086, + "learning_rate": 0.0001587685364281109, + "loss": 0.5271, + "step": 39650 + }, + { + "epoch": 0.6186435390278905, + "grad_norm": 1.5581507682800293, + "learning_rate": 0.00015875813730995613, + "loss": 0.2482, + "step": 39660 + }, + { + "epoch": 0.6187995258002121, + "grad_norm": 1.7416914701461792, + "learning_rate": 0.00015874773819180132, + "loss": 0.2018, + "step": 39670 + }, + { + "epoch": 0.6189555125725339, + "grad_norm": 0.6761413216590881, + "learning_rate": 0.00015873733907364657, + "loss": 0.1048, + "step": 39680 + }, + { + "epoch": 0.6191114993448555, + "grad_norm": 1.2344286441802979, + "learning_rate": 0.00015872693995549176, + "loss": 0.1492, + "step": 39690 + }, + { + "epoch": 0.6192674861171773, + "grad_norm": 4.924708843231201, + "learning_rate": 0.000158716540837337, + "loss": 0.2234, + "step": 39700 + }, + { + "epoch": 0.619423472889499, + "grad_norm": 2.3538808822631836, + "learning_rate": 0.0001587061417191822, + "loss": 0.3405, + "step": 39710 + }, + { + "epoch": 0.6195794596618207, + "grad_norm": 0.24888922274112701, + "learning_rate": 0.00015869574260102745, + "loss": 0.1021, + "step": 39720 + }, + { + "epoch": 0.6197354464341424, + "grad_norm": 1.2719457149505615, + "learning_rate": 0.00015868534348287264, + "loss": 0.2832, + "step": 39730 + }, + { + "epoch": 0.6198914332064641, + "grad_norm": 2.80794358253479, + "learning_rate": 0.0001586749443647179, + "loss": 0.2295, + "step": 39740 + }, + { + "epoch": 0.6200474199787858, + "grad_norm": 0.09828778356313705, + "learning_rate": 0.00015866454524656308, + "loss": 0.3574, + "step": 39750 + }, + { + "epoch": 0.6202034067511075, + "grad_norm": 1.0614128112792969, + "learning_rate": 0.00015865414612840833, + "loss": 0.1886, + "step": 39760 + }, + { + "epoch": 0.6203593935234292, + "grad_norm": 0.6939805150032043, + "learning_rate": 0.00015864374701025352, + "loss": 0.3412, + "step": 39770 + }, + { + "epoch": 0.6205153802957509, + "grad_norm": 0.06902116537094116, + "learning_rate": 0.00015863334789209877, + "loss": 0.1647, + "step": 39780 + }, + { + "epoch": 0.6206713670680726, + "grad_norm": 2.05729079246521, + "learning_rate": 0.00015862294877394396, + "loss": 0.168, + "step": 39790 + }, + { + "epoch": 0.6208273538403943, + "grad_norm": 4.69445276260376, + "learning_rate": 0.0001586125496557892, + "loss": 0.1701, + "step": 39800 + }, + { + "epoch": 0.6209833406127161, + "grad_norm": 0.6905967593193054, + "learning_rate": 0.0001586021505376344, + "loss": 0.2823, + "step": 39810 + }, + { + "epoch": 0.6211393273850377, + "grad_norm": 3.0361857414245605, + "learning_rate": 0.00015859175141947964, + "loss": 0.1839, + "step": 39820 + }, + { + "epoch": 0.6212953141573595, + "grad_norm": 0.08303021639585495, + "learning_rate": 0.00015858135230132484, + "loss": 0.2183, + "step": 39830 + }, + { + "epoch": 0.6214513009296811, + "grad_norm": 3.230077028274536, + "learning_rate": 0.00015857095318317008, + "loss": 0.1727, + "step": 39840 + }, + { + "epoch": 0.6216072877020029, + "grad_norm": 1.2884387969970703, + "learning_rate": 0.00015856055406501528, + "loss": 0.1081, + "step": 39850 + }, + { + "epoch": 0.6217632744743246, + "grad_norm": 2.912475109100342, + "learning_rate": 0.00015855015494686052, + "loss": 0.2695, + "step": 39860 + }, + { + "epoch": 0.6219192612466463, + "grad_norm": 0.14122223854064941, + "learning_rate": 0.00015853975582870572, + "loss": 0.3006, + "step": 39870 + }, + { + "epoch": 0.622075248018968, + "grad_norm": 0.8160560131072998, + "learning_rate": 0.00015852935671055096, + "loss": 0.1204, + "step": 39880 + }, + { + "epoch": 0.6222312347912897, + "grad_norm": 1.6930943727493286, + "learning_rate": 0.00015851895759239615, + "loss": 0.1809, + "step": 39890 + }, + { + "epoch": 0.6223872215636114, + "grad_norm": 0.008987984620034695, + "learning_rate": 0.0001585085584742414, + "loss": 0.125, + "step": 39900 + }, + { + "epoch": 0.6225432083359331, + "grad_norm": 2.6068177223205566, + "learning_rate": 0.0001584981593560866, + "loss": 0.3131, + "step": 39910 + }, + { + "epoch": 0.6226991951082548, + "grad_norm": 0.5021041035652161, + "learning_rate": 0.00015848776023793184, + "loss": 0.2023, + "step": 39920 + }, + { + "epoch": 0.6228551818805765, + "grad_norm": 1.8558531999588013, + "learning_rate": 0.00015847736111977703, + "loss": 0.3135, + "step": 39930 + }, + { + "epoch": 0.6230111686528982, + "grad_norm": 0.019155239686369896, + "learning_rate": 0.00015846696200162228, + "loss": 0.2282, + "step": 39940 + }, + { + "epoch": 0.6231671554252199, + "grad_norm": 0.47993865609169006, + "learning_rate": 0.00015845656288346747, + "loss": 0.1143, + "step": 39950 + }, + { + "epoch": 0.6233231421975417, + "grad_norm": 2.553943157196045, + "learning_rate": 0.00015844616376531272, + "loss": 0.2061, + "step": 39960 + }, + { + "epoch": 0.6234791289698633, + "grad_norm": 0.22600245475769043, + "learning_rate": 0.0001584357646471579, + "loss": 0.1808, + "step": 39970 + }, + { + "epoch": 0.6236351157421851, + "grad_norm": 0.7153030037879944, + "learning_rate": 0.00015842536552900316, + "loss": 0.2221, + "step": 39980 + }, + { + "epoch": 0.6237911025145068, + "grad_norm": 0.12053301930427551, + "learning_rate": 0.00015841496641084835, + "loss": 0.2349, + "step": 39990 + }, + { + "epoch": 0.6239470892868285, + "grad_norm": 0.44607964158058167, + "learning_rate": 0.0001584045672926936, + "loss": 0.234, + "step": 40000 + }, + { + "epoch": 0.6241030760591502, + "grad_norm": 4.0730366706848145, + "learning_rate": 0.0001583941681745388, + "loss": 0.3213, + "step": 40010 + }, + { + "epoch": 0.6242590628314719, + "grad_norm": 2.8753201961517334, + "learning_rate": 0.00015838376905638404, + "loss": 0.1745, + "step": 40020 + }, + { + "epoch": 0.6244150496037936, + "grad_norm": 1.7546380758285522, + "learning_rate": 0.00015837336993822926, + "loss": 0.2564, + "step": 40030 + }, + { + "epoch": 0.6245710363761153, + "grad_norm": 2.7480337619781494, + "learning_rate": 0.00015836297082007448, + "loss": 0.2722, + "step": 40040 + }, + { + "epoch": 0.624727023148437, + "grad_norm": 1.9213544130325317, + "learning_rate": 0.0001583525717019197, + "loss": 0.3495, + "step": 40050 + }, + { + "epoch": 0.6248830099207587, + "grad_norm": 1.5569000244140625, + "learning_rate": 0.00015834217258376492, + "loss": 0.1744, + "step": 40060 + }, + { + "epoch": 0.6250389966930804, + "grad_norm": 0.4767601788043976, + "learning_rate": 0.00015833177346561014, + "loss": 0.1739, + "step": 40070 + }, + { + "epoch": 0.6251949834654021, + "grad_norm": 3.514345407485962, + "learning_rate": 0.00015832137434745536, + "loss": 0.3174, + "step": 40080 + }, + { + "epoch": 0.6253509702377238, + "grad_norm": 0.7669044137001038, + "learning_rate": 0.00015831097522930057, + "loss": 0.1893, + "step": 40090 + }, + { + "epoch": 0.6255069570100456, + "grad_norm": 0.3902212083339691, + "learning_rate": 0.0001583005761111458, + "loss": 0.2533, + "step": 40100 + }, + { + "epoch": 0.6256629437823673, + "grad_norm": 1.3150569200515747, + "learning_rate": 0.00015829017699299101, + "loss": 0.2491, + "step": 40110 + }, + { + "epoch": 0.625818930554689, + "grad_norm": 0.9833875894546509, + "learning_rate": 0.00015827977787483623, + "loss": 0.2889, + "step": 40120 + }, + { + "epoch": 0.6259749173270107, + "grad_norm": 1.0952039957046509, + "learning_rate": 0.00015826937875668145, + "loss": 0.2213, + "step": 40130 + }, + { + "epoch": 0.6261309040993324, + "grad_norm": 0.32410308718681335, + "learning_rate": 0.00015825897963852665, + "loss": 0.1123, + "step": 40140 + }, + { + "epoch": 0.6262868908716541, + "grad_norm": 0.05873742699623108, + "learning_rate": 0.0001582485805203719, + "loss": 0.1944, + "step": 40150 + }, + { + "epoch": 0.6264428776439758, + "grad_norm": 0.936841607093811, + "learning_rate": 0.00015823818140221709, + "loss": 0.2605, + "step": 40160 + }, + { + "epoch": 0.6265988644162975, + "grad_norm": 3.288151264190674, + "learning_rate": 0.00015822778228406233, + "loss": 0.201, + "step": 40170 + }, + { + "epoch": 0.6267548511886192, + "grad_norm": 0.45985645055770874, + "learning_rate": 0.00015821738316590752, + "loss": 0.094, + "step": 40180 + }, + { + "epoch": 0.6269108379609409, + "grad_norm": 3.8212318420410156, + "learning_rate": 0.00015820698404775277, + "loss": 0.2554, + "step": 40190 + }, + { + "epoch": 0.6270668247332626, + "grad_norm": 1.8019777536392212, + "learning_rate": 0.00015819658492959796, + "loss": 0.5479, + "step": 40200 + }, + { + "epoch": 0.6272228115055843, + "grad_norm": 0.49328944087028503, + "learning_rate": 0.0001581861858114432, + "loss": 0.2608, + "step": 40210 + }, + { + "epoch": 0.627378798277906, + "grad_norm": 0.30712059140205383, + "learning_rate": 0.0001581757866932884, + "loss": 0.235, + "step": 40220 + }, + { + "epoch": 0.6275347850502278, + "grad_norm": 3.460827112197876, + "learning_rate": 0.00015816538757513365, + "loss": 0.3527, + "step": 40230 + }, + { + "epoch": 0.6276907718225494, + "grad_norm": 1.6247568130493164, + "learning_rate": 0.00015815498845697884, + "loss": 0.3031, + "step": 40240 + }, + { + "epoch": 0.6278467585948712, + "grad_norm": 0.5068777203559875, + "learning_rate": 0.0001581445893388241, + "loss": 0.3529, + "step": 40250 + }, + { + "epoch": 0.6280027453671929, + "grad_norm": 0.2704099416732788, + "learning_rate": 0.00015813419022066928, + "loss": 0.1043, + "step": 40260 + }, + { + "epoch": 0.6281587321395146, + "grad_norm": 0.07768986374139786, + "learning_rate": 0.00015812379110251453, + "loss": 0.257, + "step": 40270 + }, + { + "epoch": 0.6283147189118363, + "grad_norm": 2.6940250396728516, + "learning_rate": 0.00015811339198435972, + "loss": 0.1694, + "step": 40280 + }, + { + "epoch": 0.628470705684158, + "grad_norm": 2.3470752239227295, + "learning_rate": 0.00015810299286620497, + "loss": 0.2196, + "step": 40290 + }, + { + "epoch": 0.6286266924564797, + "grad_norm": 1.6333197355270386, + "learning_rate": 0.00015809259374805016, + "loss": 0.2452, + "step": 40300 + }, + { + "epoch": 0.6287826792288014, + "grad_norm": 0.28753232955932617, + "learning_rate": 0.0001580821946298954, + "loss": 0.0601, + "step": 40310 + }, + { + "epoch": 0.6289386660011231, + "grad_norm": 0.8923532366752625, + "learning_rate": 0.0001580717955117406, + "loss": 0.455, + "step": 40320 + }, + { + "epoch": 0.6290946527734448, + "grad_norm": 2.404088020324707, + "learning_rate": 0.00015806139639358585, + "loss": 0.466, + "step": 40330 + }, + { + "epoch": 0.6292506395457665, + "grad_norm": 1.0499215126037598, + "learning_rate": 0.00015805099727543104, + "loss": 0.4372, + "step": 40340 + }, + { + "epoch": 0.6294066263180882, + "grad_norm": 0.21907389163970947, + "learning_rate": 0.00015804059815727629, + "loss": 0.1628, + "step": 40350 + }, + { + "epoch": 0.62956261309041, + "grad_norm": 1.1540305614471436, + "learning_rate": 0.00015803019903912148, + "loss": 0.1888, + "step": 40360 + }, + { + "epoch": 0.6297185998627316, + "grad_norm": 0.44641104340553284, + "learning_rate": 0.00015801979992096672, + "loss": 0.2372, + "step": 40370 + }, + { + "epoch": 0.6298745866350534, + "grad_norm": 1.5536185503005981, + "learning_rate": 0.00015800940080281192, + "loss": 0.0855, + "step": 40380 + }, + { + "epoch": 0.630030573407375, + "grad_norm": 5.066720485687256, + "learning_rate": 0.00015799900168465716, + "loss": 0.3324, + "step": 40390 + }, + { + "epoch": 0.6301865601796968, + "grad_norm": 1.0257264375686646, + "learning_rate": 0.00015798860256650236, + "loss": 0.1872, + "step": 40400 + }, + { + "epoch": 0.6303425469520185, + "grad_norm": 1.108614206314087, + "learning_rate": 0.0001579782034483476, + "loss": 0.3244, + "step": 40410 + }, + { + "epoch": 0.6304985337243402, + "grad_norm": 0.9430763125419617, + "learning_rate": 0.0001579678043301928, + "loss": 0.2388, + "step": 40420 + }, + { + "epoch": 0.6306545204966619, + "grad_norm": 0.3578563332557678, + "learning_rate": 0.00015795740521203804, + "loss": 0.3072, + "step": 40430 + }, + { + "epoch": 0.6308105072689836, + "grad_norm": 0.6251798272132874, + "learning_rate": 0.00015794700609388323, + "loss": 0.3246, + "step": 40440 + }, + { + "epoch": 0.6309664940413053, + "grad_norm": 1.5441378355026245, + "learning_rate": 0.00015793660697572848, + "loss": 0.1822, + "step": 40450 + }, + { + "epoch": 0.631122480813627, + "grad_norm": 1.3051613569259644, + "learning_rate": 0.00015792620785757367, + "loss": 0.3554, + "step": 40460 + }, + { + "epoch": 0.6312784675859487, + "grad_norm": 4.274336338043213, + "learning_rate": 0.00015791580873941892, + "loss": 0.2423, + "step": 40470 + }, + { + "epoch": 0.6314344543582704, + "grad_norm": 1.4855650663375854, + "learning_rate": 0.0001579054096212641, + "loss": 0.2158, + "step": 40480 + }, + { + "epoch": 0.6315904411305922, + "grad_norm": 0.26689472794532776, + "learning_rate": 0.00015789501050310936, + "loss": 0.3227, + "step": 40490 + }, + { + "epoch": 0.6317464279029138, + "grad_norm": 2.0525012016296387, + "learning_rate": 0.00015788461138495455, + "loss": 0.2475, + "step": 40500 + }, + { + "epoch": 0.6319024146752356, + "grad_norm": 1.1261099576950073, + "learning_rate": 0.0001578742122667998, + "loss": 0.2043, + "step": 40510 + }, + { + "epoch": 0.6320584014475572, + "grad_norm": 3.018596649169922, + "learning_rate": 0.000157863813148645, + "loss": 0.2309, + "step": 40520 + }, + { + "epoch": 0.632214388219879, + "grad_norm": 2.6361162662506104, + "learning_rate": 0.00015785341403049024, + "loss": 0.2583, + "step": 40530 + }, + { + "epoch": 0.6323703749922006, + "grad_norm": 1.5715960264205933, + "learning_rate": 0.00015784301491233543, + "loss": 0.2937, + "step": 40540 + }, + { + "epoch": 0.6325263617645224, + "grad_norm": 0.03241211175918579, + "learning_rate": 0.00015783261579418068, + "loss": 0.4566, + "step": 40550 + }, + { + "epoch": 0.6326823485368441, + "grad_norm": 0.5787049531936646, + "learning_rate": 0.00015782221667602587, + "loss": 0.0956, + "step": 40560 + }, + { + "epoch": 0.6328383353091658, + "grad_norm": 2.3547847270965576, + "learning_rate": 0.00015781181755787112, + "loss": 0.1877, + "step": 40570 + }, + { + "epoch": 0.6329943220814875, + "grad_norm": 2.2037112712860107, + "learning_rate": 0.0001578014184397163, + "loss": 0.0953, + "step": 40580 + }, + { + "epoch": 0.6331503088538092, + "grad_norm": 1.3263297080993652, + "learning_rate": 0.00015779101932156153, + "loss": 0.1661, + "step": 40590 + }, + { + "epoch": 0.6333062956261309, + "grad_norm": 0.20338000357151031, + "learning_rate": 0.00015778062020340675, + "loss": 0.1421, + "step": 40600 + }, + { + "epoch": 0.6334622823984526, + "grad_norm": 1.156614065170288, + "learning_rate": 0.00015777022108525197, + "loss": 0.1927, + "step": 40610 + }, + { + "epoch": 0.6336182691707744, + "grad_norm": 4.7294416427612305, + "learning_rate": 0.0001577598219670972, + "loss": 0.2359, + "step": 40620 + }, + { + "epoch": 0.633774255943096, + "grad_norm": 0.07647063583135605, + "learning_rate": 0.0001577494228489424, + "loss": 0.2377, + "step": 40630 + }, + { + "epoch": 0.6339302427154178, + "grad_norm": 1.5807030200958252, + "learning_rate": 0.00015773902373078763, + "loss": 0.1148, + "step": 40640 + }, + { + "epoch": 0.6340862294877394, + "grad_norm": 0.17653131484985352, + "learning_rate": 0.00015772862461263285, + "loss": 0.3267, + "step": 40650 + }, + { + "epoch": 0.6342422162600612, + "grad_norm": 2.8481879234313965, + "learning_rate": 0.00015771822549447807, + "loss": 0.2792, + "step": 40660 + }, + { + "epoch": 0.6343982030323828, + "grad_norm": 1.6328898668289185, + "learning_rate": 0.0001577078263763233, + "loss": 0.1816, + "step": 40670 + }, + { + "epoch": 0.6345541898047046, + "grad_norm": 1.9256733655929565, + "learning_rate": 0.0001576974272581685, + "loss": 0.5716, + "step": 40680 + }, + { + "epoch": 0.6347101765770262, + "grad_norm": 2.0054070949554443, + "learning_rate": 0.00015768702814001373, + "loss": 0.2331, + "step": 40690 + }, + { + "epoch": 0.634866163349348, + "grad_norm": 0.592220664024353, + "learning_rate": 0.00015767662902185895, + "loss": 0.1639, + "step": 40700 + }, + { + "epoch": 0.6350221501216697, + "grad_norm": 0.7444786429405212, + "learning_rate": 0.00015766622990370417, + "loss": 0.1833, + "step": 40710 + }, + { + "epoch": 0.6351781368939914, + "grad_norm": 1.4234321117401123, + "learning_rate": 0.00015765583078554938, + "loss": 0.4297, + "step": 40720 + }, + { + "epoch": 0.6353341236663131, + "grad_norm": 3.897197723388672, + "learning_rate": 0.0001576454316673946, + "loss": 0.2512, + "step": 40730 + }, + { + "epoch": 0.6354901104386348, + "grad_norm": 0.4277057647705078, + "learning_rate": 0.00015763503254923982, + "loss": 0.2606, + "step": 40740 + }, + { + "epoch": 0.6356460972109566, + "grad_norm": 3.6413636207580566, + "learning_rate": 0.00015762463343108504, + "loss": 0.1645, + "step": 40750 + }, + { + "epoch": 0.6358020839832782, + "grad_norm": 0.6908426284790039, + "learning_rate": 0.00015761423431293026, + "loss": 0.185, + "step": 40760 + }, + { + "epoch": 0.6359580707556, + "grad_norm": 0.7075207829475403, + "learning_rate": 0.00015760383519477548, + "loss": 0.2374, + "step": 40770 + }, + { + "epoch": 0.6361140575279216, + "grad_norm": 10.855485916137695, + "learning_rate": 0.0001575934360766207, + "loss": 0.2817, + "step": 40780 + }, + { + "epoch": 0.6362700443002434, + "grad_norm": 1.455926537513733, + "learning_rate": 0.00015758303695846592, + "loss": 0.1535, + "step": 40790 + }, + { + "epoch": 0.636426031072565, + "grad_norm": 0.469672292470932, + "learning_rate": 0.00015757263784031114, + "loss": 0.2253, + "step": 40800 + }, + { + "epoch": 0.6365820178448868, + "grad_norm": 0.19487737119197845, + "learning_rate": 0.00015756223872215636, + "loss": 0.1672, + "step": 40810 + }, + { + "epoch": 0.6367380046172084, + "grad_norm": 0.022243143990635872, + "learning_rate": 0.00015755183960400158, + "loss": 0.4319, + "step": 40820 + }, + { + "epoch": 0.6368939913895302, + "grad_norm": 0.6900194883346558, + "learning_rate": 0.0001575414404858468, + "loss": 0.12, + "step": 40830 + }, + { + "epoch": 0.6370499781618518, + "grad_norm": 1.100761890411377, + "learning_rate": 0.00015753104136769202, + "loss": 0.164, + "step": 40840 + }, + { + "epoch": 0.6372059649341736, + "grad_norm": 1.8163270950317383, + "learning_rate": 0.00015752064224953724, + "loss": 0.3855, + "step": 40850 + }, + { + "epoch": 0.6373619517064953, + "grad_norm": 0.14638766646385193, + "learning_rate": 0.00015751024313138246, + "loss": 0.0791, + "step": 40860 + }, + { + "epoch": 0.637517938478817, + "grad_norm": 0.419525682926178, + "learning_rate": 0.00015749984401322768, + "loss": 0.4335, + "step": 40870 + }, + { + "epoch": 0.6376739252511388, + "grad_norm": 1.365848183631897, + "learning_rate": 0.00015748944489507293, + "loss": 0.1774, + "step": 40880 + }, + { + "epoch": 0.6378299120234604, + "grad_norm": 0.020374717190861702, + "learning_rate": 0.00015747904577691812, + "loss": 0.1826, + "step": 40890 + }, + { + "epoch": 0.6379858987957822, + "grad_norm": 3.35870623588562, + "learning_rate": 0.00015746864665876337, + "loss": 0.1956, + "step": 40900 + }, + { + "epoch": 0.6381418855681038, + "grad_norm": 0.41389143466949463, + "learning_rate": 0.00015745824754060856, + "loss": 0.304, + "step": 40910 + }, + { + "epoch": 0.6382978723404256, + "grad_norm": 4.447183609008789, + "learning_rate": 0.0001574478484224538, + "loss": 0.2884, + "step": 40920 + }, + { + "epoch": 0.6384538591127472, + "grad_norm": 0.428251713514328, + "learning_rate": 0.000157437449304299, + "loss": 0.0913, + "step": 40930 + }, + { + "epoch": 0.638609845885069, + "grad_norm": 1.440173864364624, + "learning_rate": 0.00015742705018614424, + "loss": 0.3341, + "step": 40940 + }, + { + "epoch": 0.6387658326573906, + "grad_norm": 0.37606480717658997, + "learning_rate": 0.00015741665106798944, + "loss": 0.1545, + "step": 40950 + }, + { + "epoch": 0.6389218194297124, + "grad_norm": 0.08015301823616028, + "learning_rate": 0.00015740625194983468, + "loss": 0.2598, + "step": 40960 + }, + { + "epoch": 0.639077806202034, + "grad_norm": 1.5615071058273315, + "learning_rate": 0.00015739585283167988, + "loss": 0.1958, + "step": 40970 + }, + { + "epoch": 0.6392337929743558, + "grad_norm": 0.7406660914421082, + "learning_rate": 0.00015738545371352512, + "loss": 0.2646, + "step": 40980 + }, + { + "epoch": 0.6393897797466774, + "grad_norm": 1.9117989540100098, + "learning_rate": 0.00015737505459537032, + "loss": 0.2309, + "step": 40990 + }, + { + "epoch": 0.6395457665189992, + "grad_norm": 4.485423564910889, + "learning_rate": 0.00015736465547721556, + "loss": 0.1564, + "step": 41000 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.496727351296e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41000/training_args.bin differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/added_tokens.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/added_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ea36ca9a30d42cfe00f964ed2b450595386671dc --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/added_tokens.json @@ -0,0 +1,98 @@ +{ + "": 36095, + "": 36085, + "": 36084, + "": 36083, + "": 36082, + "": 36081, + "": 36080, + "": 36079, + "": 36078, + "": 36077, + "": 36076, + "": 36094, + "": 36075, + "": 36074, + "": 36073, + "": 36072, + "": 36071, + "": 36070, + "": 36069, + "": 36068, + "": 36067, + "": 36066, + "": 36093, + "": 36065, + "": 36064, + "": 36063, + "": 36062, + "": 36061, + "": 36060, + "": 36059, + "": 36058, + "": 36057, + "": 36056, + "": 36092, + "": 36055, + "": 36054, + "": 36053, + "": 36052, + "": 36051, + "": 36050, + "": 36049, + "": 36048, + "": 36047, + "": 36046, + "": 36091, + "": 36045, + "": 36044, + "": 36043, + "": 36042, + "": 36041, + "": 36040, + "": 36039, + "": 36038, + "": 36037, + "": 36036, + "": 36090, + "": 36035, + "": 36034, + "": 36033, + "": 36032, + "": 36031, + "": 36030, + "": 36029, + "": 36028, + "": 36027, + "": 36026, + "": 36089, + "": 36025, + "": 36024, + "": 36023, + "": 36022, + "": 36021, + "": 36020, + "": 36019, + "": 36018, + "": 36017, + "": 36016, + "": 36088, + "": 36015, + "": 36014, + "": 36013, + "": 36012, + "": 36011, + "": 36010, + "": 36009, + "": 36008, + "": 36007, + "": 36006, + "": 36087, + "": 36005, + "": 36004, + "": 36003, + "": 36002, + "": 36001, + "": 36000, + "": 36086 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/config.json new file mode 100644 index 0000000000000000000000000000000000000000..402574816dbd44d7c00b58882614b0cba909b7da --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/config.json @@ -0,0 +1,31 @@ +{ + "architectures": [ + "T5ForConditionalGeneration" + ], + "classifier_dropout": 0.0, + "d_ff": 3072, + "d_kv": 64, + "d_model": 768, + "decoder_start_token_id": 0, + "dense_act_fn": "relu", + "dropout_rate": 0.1, + "eos_token_id": 1, + "feed_forward_proj": "relu", + "initializer_factor": 1.0, + "is_encoder_decoder": true, + "is_gated_act": false, + "layer_norm_epsilon": 1e-06, + "model_type": "t5", + "n_positions": 512, + "num_decoder_layers": 12, + "num_heads": 12, + "num_layers": 12, + "output_past": true, + "pad_token_id": 0, + "relative_attention_max_distance": 128, + "relative_attention_num_buckets": 32, + "torch_dtype": "float32", + "transformers_version": "4.53.2", + "use_cache": true, + "vocab_size": 36096 +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/generation_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..88a411aa38dffcace543088cff3153003de6e01a --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "decoder_start_token_id": 0, + "eos_token_id": 1, + "pad_token_id": 0, + "transformers_version": "4.53.2" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/model.safetensors b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f7b4a1bfbac3ca1f1a8b1db8a7d7d0c8c579bb88 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7ef000c5312bc374266e7da5d7aa6c5fcd7b48a68806aba1276b57fa5e4b1180 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/optimizer.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..d5e41dd8a48a2fe53b347eeb428e9fa41e3b9154 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb5d3f3b99990858297e2d9cc375a9dc19764fccbcf0b511bb12fadbc3ccf699 +size 62652416 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/rng_state.pth b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..114c4d5534de94af6ff5a031126e47e34c5c40e2 Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/rng_state.pth differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/scheduler.pt b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..71778ecd71973b16287d0a2cddde56b74a8f16ca Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/scheduler.pt differ diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/special_tokens_map.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..843b3344f47b1783c48b5ac91bb6015ae9d3c4be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/special_tokens_map.json @@ -0,0 +1,121 @@ +{ + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/spiece.model b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/spiece.model new file mode 100644 index 0000000000000000000000000000000000000000..f8bddaf892bdf23d2148f3a3b358f16c5c45c7be --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/spiece.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6 +size 820370 diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/tokenizer_config.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7da9aea82b39809d9fbe6214e6f4fa2340b695f8 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/tokenizer_config.json @@ -0,0 +1,905 @@ +{ + "add_prefix_space": true, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "36000": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36001": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36002": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36003": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36004": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36005": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36006": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36007": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36008": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36009": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36010": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36011": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36012": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36013": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36014": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36015": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36016": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36017": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36018": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36019": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36020": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36021": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36022": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36023": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36024": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36025": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36026": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36027": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36028": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36029": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36030": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36031": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36032": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36033": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36034": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36035": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36036": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36037": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36038": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36039": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36040": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36041": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36042": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36043": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36044": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36045": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36046": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36047": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36048": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36049": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36050": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36051": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36052": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36053": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36054": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36055": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36056": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36057": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36058": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36059": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36060": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36061": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36062": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36063": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36064": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36065": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36066": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36067": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36068": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36069": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36070": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36071": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36072": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36073": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36074": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36075": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36076": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36077": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36078": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36079": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36080": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36081": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36082": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36083": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36084": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36085": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36086": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36087": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36088": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36089": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36090": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36091": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36092": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36093": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36094": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + }, + "36095": { + "content": "", + "lstrip": true, + "normalized": false, + "rstrip": true, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "", + "" + ], + "clean_up_tokenization_spaces": false, + "eos_token": "", + "extra_ids": 96, + "extra_special_tokens": {}, + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "tokenizer_class": "T5Tokenizer", + "unk_token": "" +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/trainer_state.json b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..0f574e370b60f74e0431d7e7feefbf923ae3c752 --- /dev/null +++ b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/trainer_state.json @@ -0,0 +1,29084 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.6473451051350846, + "eval_steps": 500, + "global_step": 41500, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00015598677232170712, + "grad_norm": 3.216968059539795, + "learning_rate": 0.0001999906407936607, + "loss": 4.7744, + "step": 10 + }, + { + "epoch": 0.00031197354464341425, + "grad_norm": 9.366360664367676, + "learning_rate": 0.00019998024167550594, + "loss": 0.9639, + "step": 20 + }, + { + "epoch": 0.00046796031696512137, + "grad_norm": 4.382473945617676, + "learning_rate": 0.00019996984255735113, + "loss": 0.6338, + "step": 30 + }, + { + "epoch": 0.0006239470892868285, + "grad_norm": 14.415914535522461, + "learning_rate": 0.00019995944343919637, + "loss": 0.4962, + "step": 40 + }, + { + "epoch": 0.0007799338616085356, + "grad_norm": 7.440260410308838, + "learning_rate": 0.00019994904432104157, + "loss": 0.4619, + "step": 50 + }, + { + "epoch": 0.0009359206339302427, + "grad_norm": 27.18022918701172, + "learning_rate": 0.00019993864520288681, + "loss": 0.5335, + "step": 60 + }, + { + "epoch": 0.0010919074062519499, + "grad_norm": 7.954453468322754, + "learning_rate": 0.000199928246084732, + "loss": 0.8085, + "step": 70 + }, + { + "epoch": 0.001247894178573657, + "grad_norm": 4.762608528137207, + "learning_rate": 0.00019991784696657725, + "loss": 0.6333, + "step": 80 + }, + { + "epoch": 0.0014038809508953641, + "grad_norm": 9.577717781066895, + "learning_rate": 0.00019990744784842245, + "loss": 0.4558, + "step": 90 + }, + { + "epoch": 0.0015598677232170712, + "grad_norm": 7.586806297302246, + "learning_rate": 0.0001998970487302677, + "loss": 0.3227, + "step": 100 + }, + { + "epoch": 0.0017158544955387784, + "grad_norm": 4.262249946594238, + "learning_rate": 0.00019988664961211288, + "loss": 0.3261, + "step": 110 + }, + { + "epoch": 0.0018718412678604855, + "grad_norm": 6.639514923095703, + "learning_rate": 0.00019987625049395813, + "loss": 0.6788, + "step": 120 + }, + { + "epoch": 0.0020278280401821924, + "grad_norm": 3.7448127269744873, + "learning_rate": 0.00019986585137580332, + "loss": 0.4224, + "step": 130 + }, + { + "epoch": 0.0021838148125038997, + "grad_norm": 13.205582618713379, + "learning_rate": 0.00019985545225764857, + "loss": 0.4074, + "step": 140 + }, + { + "epoch": 0.0023398015848256066, + "grad_norm": 5.55623722076416, + "learning_rate": 0.00019984505313949376, + "loss": 0.3408, + "step": 150 + }, + { + "epoch": 0.002495788357147314, + "grad_norm": 10.253294944763184, + "learning_rate": 0.000199834654021339, + "loss": 0.4513, + "step": 160 + }, + { + "epoch": 0.002651775129469021, + "grad_norm": 12.778948783874512, + "learning_rate": 0.0001998242549031842, + "loss": 0.4287, + "step": 170 + }, + { + "epoch": 0.0028077619017907282, + "grad_norm": 8.961956977844238, + "learning_rate": 0.00019981385578502945, + "loss": 0.6746, + "step": 180 + }, + { + "epoch": 0.002963748674112435, + "grad_norm": 6.611206531524658, + "learning_rate": 0.00019980345666687467, + "loss": 0.4106, + "step": 190 + }, + { + "epoch": 0.0031197354464341425, + "grad_norm": 4.179671764373779, + "learning_rate": 0.0001997930575487199, + "loss": 0.4234, + "step": 200 + }, + { + "epoch": 0.0032757222187558494, + "grad_norm": 7.957104206085205, + "learning_rate": 0.0001997826584305651, + "loss": 0.4062, + "step": 210 + }, + { + "epoch": 0.0034317089910775567, + "grad_norm": 10.04617977142334, + "learning_rate": 0.00019977225931241033, + "loss": 0.4792, + "step": 220 + }, + { + "epoch": 0.0035876957633992636, + "grad_norm": 5.119971752166748, + "learning_rate": 0.00019976186019425555, + "loss": 0.3385, + "step": 230 + }, + { + "epoch": 0.003743682535720971, + "grad_norm": 6.039366722106934, + "learning_rate": 0.00019975146107610077, + "loss": 0.4406, + "step": 240 + }, + { + "epoch": 0.003899669308042678, + "grad_norm": 9.345914840698242, + "learning_rate": 0.000199741061957946, + "loss": 0.5163, + "step": 250 + }, + { + "epoch": 0.004055656080364385, + "grad_norm": 3.707460641860962, + "learning_rate": 0.0001997306628397912, + "loss": 0.2971, + "step": 260 + }, + { + "epoch": 0.0042116428526860926, + "grad_norm": 11.500152587890625, + "learning_rate": 0.00019972026372163643, + "loss": 0.3659, + "step": 270 + }, + { + "epoch": 0.0043676296250077995, + "grad_norm": 0.7893226742744446, + "learning_rate": 0.00019970986460348165, + "loss": 0.5052, + "step": 280 + }, + { + "epoch": 0.004523616397329506, + "grad_norm": 3.2410888671875, + "learning_rate": 0.00019969946548532687, + "loss": 0.4038, + "step": 290 + }, + { + "epoch": 0.004679603169651213, + "grad_norm": 4.468445777893066, + "learning_rate": 0.00019968906636717208, + "loss": 0.6076, + "step": 300 + }, + { + "epoch": 0.004835589941972921, + "grad_norm": 8.308616638183594, + "learning_rate": 0.0001996786672490173, + "loss": 0.4465, + "step": 310 + }, + { + "epoch": 0.004991576714294628, + "grad_norm": 3.736629009246826, + "learning_rate": 0.0001996682681308625, + "loss": 0.4133, + "step": 320 + }, + { + "epoch": 0.005147563486616335, + "grad_norm": 2.157092809677124, + "learning_rate": 0.00019965786901270774, + "loss": 0.3663, + "step": 330 + }, + { + "epoch": 0.005303550258938042, + "grad_norm": 5.357859134674072, + "learning_rate": 0.00019964746989455294, + "loss": 0.5599, + "step": 340 + }, + { + "epoch": 0.0054595370312597495, + "grad_norm": 5.352982044219971, + "learning_rate": 0.00019963707077639818, + "loss": 0.6796, + "step": 350 + }, + { + "epoch": 0.0056155238035814565, + "grad_norm": 3.7645204067230225, + "learning_rate": 0.00019962667165824338, + "loss": 0.5774, + "step": 360 + }, + { + "epoch": 0.005771510575903163, + "grad_norm": 6.222738265991211, + "learning_rate": 0.00019961627254008862, + "loss": 0.641, + "step": 370 + }, + { + "epoch": 0.00592749734822487, + "grad_norm": 6.460178852081299, + "learning_rate": 0.00019960587342193381, + "loss": 0.3702, + "step": 380 + }, + { + "epoch": 0.006083484120546578, + "grad_norm": 3.1196858882904053, + "learning_rate": 0.00019959547430377906, + "loss": 0.5036, + "step": 390 + }, + { + "epoch": 0.006239470892868285, + "grad_norm": 7.03672981262207, + "learning_rate": 0.00019958507518562425, + "loss": 0.41, + "step": 400 + }, + { + "epoch": 0.006395457665189992, + "grad_norm": 15.546775817871094, + "learning_rate": 0.0001995746760674695, + "loss": 0.5623, + "step": 410 + }, + { + "epoch": 0.006551444437511699, + "grad_norm": 4.868813514709473, + "learning_rate": 0.0001995642769493147, + "loss": 0.4, + "step": 420 + }, + { + "epoch": 0.0067074312098334065, + "grad_norm": 3.5767409801483154, + "learning_rate": 0.00019955387783115994, + "loss": 0.5193, + "step": 430 + }, + { + "epoch": 0.0068634179821551134, + "grad_norm": 10.07199764251709, + "learning_rate": 0.00019954347871300513, + "loss": 0.5692, + "step": 440 + }, + { + "epoch": 0.00701940475447682, + "grad_norm": 2.809025764465332, + "learning_rate": 0.00019953307959485038, + "loss": 0.4361, + "step": 450 + }, + { + "epoch": 0.007175391526798527, + "grad_norm": 14.244531631469727, + "learning_rate": 0.00019952268047669557, + "loss": 0.3646, + "step": 460 + }, + { + "epoch": 0.007331378299120235, + "grad_norm": 5.827071666717529, + "learning_rate": 0.00019951228135854082, + "loss": 0.5302, + "step": 470 + }, + { + "epoch": 0.007487365071441942, + "grad_norm": 9.067925453186035, + "learning_rate": 0.000199501882240386, + "loss": 0.5045, + "step": 480 + }, + { + "epoch": 0.007643351843763649, + "grad_norm": 8.674382209777832, + "learning_rate": 0.00019949148312223126, + "loss": 0.5778, + "step": 490 + }, + { + "epoch": 0.007799338616085356, + "grad_norm": 8.94734001159668, + "learning_rate": 0.00019948108400407645, + "loss": 0.4141, + "step": 500 + }, + { + "epoch": 0.007955325388407063, + "grad_norm": 10.810057640075684, + "learning_rate": 0.0001994706848859217, + "loss": 0.4069, + "step": 510 + }, + { + "epoch": 0.00811131216072877, + "grad_norm": 9.531766891479492, + "learning_rate": 0.0001994602857677669, + "loss": 0.5935, + "step": 520 + }, + { + "epoch": 0.008267298933050476, + "grad_norm": 13.168464660644531, + "learning_rate": 0.00019944988664961214, + "loss": 0.4725, + "step": 530 + }, + { + "epoch": 0.008423285705372185, + "grad_norm": 4.697941780090332, + "learning_rate": 0.00019943948753145733, + "loss": 0.4516, + "step": 540 + }, + { + "epoch": 0.008579272477693892, + "grad_norm": 2.92124342918396, + "learning_rate": 0.00019942908841330258, + "loss": 0.4562, + "step": 550 + }, + { + "epoch": 0.008735259250015599, + "grad_norm": 9.23039722442627, + "learning_rate": 0.00019941868929514777, + "loss": 0.4051, + "step": 560 + }, + { + "epoch": 0.008891246022337306, + "grad_norm": 10.528753280639648, + "learning_rate": 0.00019940829017699302, + "loss": 0.3475, + "step": 570 + }, + { + "epoch": 0.009047232794659013, + "grad_norm": 6.963693618774414, + "learning_rate": 0.0001993978910588382, + "loss": 0.5338, + "step": 580 + }, + { + "epoch": 0.00920321956698072, + "grad_norm": 6.09390926361084, + "learning_rate": 0.00019938749194068345, + "loss": 0.6501, + "step": 590 + }, + { + "epoch": 0.009359206339302427, + "grad_norm": 4.319295406341553, + "learning_rate": 0.00019937709282252865, + "loss": 0.4163, + "step": 600 + }, + { + "epoch": 0.009515193111624133, + "grad_norm": 7.514628887176514, + "learning_rate": 0.0001993666937043739, + "loss": 0.5193, + "step": 610 + }, + { + "epoch": 0.009671179883945842, + "grad_norm": 5.011287212371826, + "learning_rate": 0.00019935629458621909, + "loss": 0.5743, + "step": 620 + }, + { + "epoch": 0.009827166656267549, + "grad_norm": 6.090816974639893, + "learning_rate": 0.00019934589546806433, + "loss": 0.3902, + "step": 630 + }, + { + "epoch": 0.009983153428589256, + "grad_norm": 10.161015510559082, + "learning_rate": 0.00019933549634990953, + "loss": 0.4463, + "step": 640 + }, + { + "epoch": 0.010139140200910963, + "grad_norm": 6.14881706237793, + "learning_rate": 0.00019932509723175477, + "loss": 0.5009, + "step": 650 + }, + { + "epoch": 0.01029512697323267, + "grad_norm": 3.2960498332977295, + "learning_rate": 0.00019931469811359996, + "loss": 0.4597, + "step": 660 + }, + { + "epoch": 0.010451113745554377, + "grad_norm": 7.595890998840332, + "learning_rate": 0.0001993042989954452, + "loss": 0.5948, + "step": 670 + }, + { + "epoch": 0.010607100517876084, + "grad_norm": 7.178264617919922, + "learning_rate": 0.0001992938998772904, + "loss": 0.7185, + "step": 680 + }, + { + "epoch": 0.01076308729019779, + "grad_norm": 4.092066764831543, + "learning_rate": 0.00019928350075913565, + "loss": 0.3544, + "step": 690 + }, + { + "epoch": 0.010919074062519499, + "grad_norm": 13.203527450561523, + "learning_rate": 0.00019927310164098084, + "loss": 0.5719, + "step": 700 + }, + { + "epoch": 0.011075060834841206, + "grad_norm": 5.39072847366333, + "learning_rate": 0.0001992627025228261, + "loss": 0.5527, + "step": 710 + }, + { + "epoch": 0.011231047607162913, + "grad_norm": 5.769185543060303, + "learning_rate": 0.00019925230340467128, + "loss": 0.4555, + "step": 720 + }, + { + "epoch": 0.01138703437948462, + "grad_norm": 3.0080807209014893, + "learning_rate": 0.00019924190428651653, + "loss": 0.277, + "step": 730 + }, + { + "epoch": 0.011543021151806327, + "grad_norm": 1.0591176748275757, + "learning_rate": 0.00019923150516836172, + "loss": 0.513, + "step": 740 + }, + { + "epoch": 0.011699007924128034, + "grad_norm": 8.458100318908691, + "learning_rate": 0.00019922110605020697, + "loss": 0.5038, + "step": 750 + }, + { + "epoch": 0.01185499469644974, + "grad_norm": 3.121551990509033, + "learning_rate": 0.00019921070693205216, + "loss": 0.2639, + "step": 760 + }, + { + "epoch": 0.012010981468771447, + "grad_norm": 6.6408610343933105, + "learning_rate": 0.00019920030781389738, + "loss": 0.6632, + "step": 770 + }, + { + "epoch": 0.012166968241093156, + "grad_norm": 7.180611610412598, + "learning_rate": 0.0001991899086957426, + "loss": 0.5173, + "step": 780 + }, + { + "epoch": 0.012322955013414863, + "grad_norm": 4.404491901397705, + "learning_rate": 0.00019917950957758782, + "loss": 0.3628, + "step": 790 + }, + { + "epoch": 0.01247894178573657, + "grad_norm": 5.945431232452393, + "learning_rate": 0.00019916911045943304, + "loss": 0.3574, + "step": 800 + }, + { + "epoch": 0.012634928558058277, + "grad_norm": 5.105412483215332, + "learning_rate": 0.00019915871134127826, + "loss": 0.7038, + "step": 810 + }, + { + "epoch": 0.012790915330379984, + "grad_norm": 3.7237210273742676, + "learning_rate": 0.00019914831222312348, + "loss": 0.4426, + "step": 820 + }, + { + "epoch": 0.01294690210270169, + "grad_norm": 10.387605667114258, + "learning_rate": 0.0001991379131049687, + "loss": 0.3152, + "step": 830 + }, + { + "epoch": 0.013102888875023398, + "grad_norm": 2.875211477279663, + "learning_rate": 0.00019912751398681392, + "loss": 0.424, + "step": 840 + }, + { + "epoch": 0.013258875647345104, + "grad_norm": 1.0501248836517334, + "learning_rate": 0.00019911711486865914, + "loss": 0.5571, + "step": 850 + }, + { + "epoch": 0.013414862419666813, + "grad_norm": 6.748145580291748, + "learning_rate": 0.00019910671575050436, + "loss": 0.5226, + "step": 860 + }, + { + "epoch": 0.01357084919198852, + "grad_norm": 6.29374885559082, + "learning_rate": 0.00019909631663234958, + "loss": 0.3036, + "step": 870 + }, + { + "epoch": 0.013726835964310227, + "grad_norm": 5.8310699462890625, + "learning_rate": 0.0001990859175141948, + "loss": 0.4704, + "step": 880 + }, + { + "epoch": 0.013882822736631934, + "grad_norm": 1.4619086980819702, + "learning_rate": 0.00019907551839604002, + "loss": 0.3639, + "step": 890 + }, + { + "epoch": 0.01403880950895364, + "grad_norm": 8.177449226379395, + "learning_rate": 0.00019906511927788524, + "loss": 0.3853, + "step": 900 + }, + { + "epoch": 0.014194796281275348, + "grad_norm": 2.045549154281616, + "learning_rate": 0.00019905472015973046, + "loss": 0.3728, + "step": 910 + }, + { + "epoch": 0.014350783053597055, + "grad_norm": 6.4138875007629395, + "learning_rate": 0.00019904432104157568, + "loss": 0.4615, + "step": 920 + }, + { + "epoch": 0.014506769825918761, + "grad_norm": 2.442739486694336, + "learning_rate": 0.0001990339219234209, + "loss": 0.3592, + "step": 930 + }, + { + "epoch": 0.01466275659824047, + "grad_norm": 3.0741333961486816, + "learning_rate": 0.00019902352280526611, + "loss": 0.4118, + "step": 940 + }, + { + "epoch": 0.014818743370562177, + "grad_norm": 5.163224220275879, + "learning_rate": 0.00019901312368711133, + "loss": 0.7433, + "step": 950 + }, + { + "epoch": 0.014974730142883884, + "grad_norm": 5.6979193687438965, + "learning_rate": 0.00019900272456895655, + "loss": 0.3094, + "step": 960 + }, + { + "epoch": 0.01513071691520559, + "grad_norm": 3.2027347087860107, + "learning_rate": 0.00019899232545080177, + "loss": 0.3749, + "step": 970 + }, + { + "epoch": 0.015286703687527298, + "grad_norm": 9.097466468811035, + "learning_rate": 0.000198981926332647, + "loss": 0.5344, + "step": 980 + }, + { + "epoch": 0.015442690459849005, + "grad_norm": 3.730250835418701, + "learning_rate": 0.0001989715272144922, + "loss": 0.4085, + "step": 990 + }, + { + "epoch": 0.015598677232170712, + "grad_norm": 4.03507661819458, + "learning_rate": 0.00019896112809633743, + "loss": 0.7935, + "step": 1000 + }, + { + "epoch": 0.01575466400449242, + "grad_norm": 6.676806449890137, + "learning_rate": 0.00019895072897818265, + "loss": 0.4276, + "step": 1010 + }, + { + "epoch": 0.015910650776814125, + "grad_norm": 2.294931173324585, + "learning_rate": 0.00019894032986002787, + "loss": 0.2544, + "step": 1020 + }, + { + "epoch": 0.016066637549135832, + "grad_norm": 15.660760879516602, + "learning_rate": 0.0001989299307418731, + "loss": 0.2441, + "step": 1030 + }, + { + "epoch": 0.01622262432145754, + "grad_norm": 3.4223055839538574, + "learning_rate": 0.00019891953162371834, + "loss": 0.4861, + "step": 1040 + }, + { + "epoch": 0.016378611093779246, + "grad_norm": 8.601675987243652, + "learning_rate": 0.00019890913250556353, + "loss": 0.7212, + "step": 1050 + }, + { + "epoch": 0.016534597866100953, + "grad_norm": 3.814286708831787, + "learning_rate": 0.00019889873338740878, + "loss": 0.4785, + "step": 1060 + }, + { + "epoch": 0.016690584638422663, + "grad_norm": 10.6279296875, + "learning_rate": 0.00019888833426925397, + "loss": 0.278, + "step": 1070 + }, + { + "epoch": 0.01684657141074437, + "grad_norm": 2.3790910243988037, + "learning_rate": 0.00019887793515109922, + "loss": 0.3589, + "step": 1080 + }, + { + "epoch": 0.017002558183066077, + "grad_norm": 8.939804077148438, + "learning_rate": 0.0001988675360329444, + "loss": 0.5794, + "step": 1090 + }, + { + "epoch": 0.017158544955387784, + "grad_norm": 7.628971576690674, + "learning_rate": 0.00019885713691478966, + "loss": 0.5317, + "step": 1100 + }, + { + "epoch": 0.01731453172770949, + "grad_norm": 1.8617459535598755, + "learning_rate": 0.00019884673779663485, + "loss": 0.433, + "step": 1110 + }, + { + "epoch": 0.017470518500031198, + "grad_norm": 6.928651809692383, + "learning_rate": 0.0001988363386784801, + "loss": 0.216, + "step": 1120 + }, + { + "epoch": 0.017626505272352905, + "grad_norm": 3.6206703186035156, + "learning_rate": 0.0001988259395603253, + "loss": 0.4963, + "step": 1130 + }, + { + "epoch": 0.01778249204467461, + "grad_norm": 4.615293025970459, + "learning_rate": 0.00019881554044217053, + "loss": 0.3579, + "step": 1140 + }, + { + "epoch": 0.01793847881699632, + "grad_norm": 3.3206489086151123, + "learning_rate": 0.00019880514132401573, + "loss": 0.3717, + "step": 1150 + }, + { + "epoch": 0.018094465589318025, + "grad_norm": 7.5789408683776855, + "learning_rate": 0.00019879474220586097, + "loss": 0.2646, + "step": 1160 + }, + { + "epoch": 0.018250452361639732, + "grad_norm": 4.467035293579102, + "learning_rate": 0.00019878434308770617, + "loss": 0.1804, + "step": 1170 + }, + { + "epoch": 0.01840643913396144, + "grad_norm": 9.318495750427246, + "learning_rate": 0.0001987739439695514, + "loss": 0.3257, + "step": 1180 + }, + { + "epoch": 0.018562425906283146, + "grad_norm": 3.1046931743621826, + "learning_rate": 0.0001987635448513966, + "loss": 0.4267, + "step": 1190 + }, + { + "epoch": 0.018718412678604853, + "grad_norm": 4.199711799621582, + "learning_rate": 0.00019875314573324185, + "loss": 0.6676, + "step": 1200 + }, + { + "epoch": 0.01887439945092656, + "grad_norm": 12.733593940734863, + "learning_rate": 0.00019874274661508705, + "loss": 0.5991, + "step": 1210 + }, + { + "epoch": 0.019030386223248267, + "grad_norm": 5.1266984939575195, + "learning_rate": 0.0001987323474969323, + "loss": 0.2945, + "step": 1220 + }, + { + "epoch": 0.019186372995569977, + "grad_norm": 6.397432804107666, + "learning_rate": 0.00019872194837877748, + "loss": 0.4774, + "step": 1230 + }, + { + "epoch": 0.019342359767891684, + "grad_norm": 3.761129379272461, + "learning_rate": 0.0001987115492606227, + "loss": 0.4345, + "step": 1240 + }, + { + "epoch": 0.01949834654021339, + "grad_norm": 2.7355475425720215, + "learning_rate": 0.00019870115014246792, + "loss": 0.497, + "step": 1250 + }, + { + "epoch": 0.019654333312535098, + "grad_norm": 2.7893755435943604, + "learning_rate": 0.00019869075102431314, + "loss": 0.2169, + "step": 1260 + }, + { + "epoch": 0.019810320084856805, + "grad_norm": 10.722111701965332, + "learning_rate": 0.00019868035190615836, + "loss": 0.3867, + "step": 1270 + }, + { + "epoch": 0.019966306857178512, + "grad_norm": 11.647889137268066, + "learning_rate": 0.00019866995278800358, + "loss": 0.4138, + "step": 1280 + }, + { + "epoch": 0.02012229362950022, + "grad_norm": 1.4313205480575562, + "learning_rate": 0.0001986595536698488, + "loss": 0.3581, + "step": 1290 + }, + { + "epoch": 0.020278280401821926, + "grad_norm": 5.193609714508057, + "learning_rate": 0.00019864915455169402, + "loss": 0.5259, + "step": 1300 + }, + { + "epoch": 0.020434267174143633, + "grad_norm": 3.1939520835876465, + "learning_rate": 0.00019863875543353924, + "loss": 0.5391, + "step": 1310 + }, + { + "epoch": 0.02059025394646534, + "grad_norm": 7.716986179351807, + "learning_rate": 0.00019862835631538446, + "loss": 0.4773, + "step": 1320 + }, + { + "epoch": 0.020746240718787046, + "grad_norm": 3.191506862640381, + "learning_rate": 0.00019861795719722968, + "loss": 0.6051, + "step": 1330 + }, + { + "epoch": 0.020902227491108753, + "grad_norm": 5.67583703994751, + "learning_rate": 0.0001986075580790749, + "loss": 0.2716, + "step": 1340 + }, + { + "epoch": 0.02105821426343046, + "grad_norm": 3.1657073497772217, + "learning_rate": 0.00019859715896092012, + "loss": 0.4838, + "step": 1350 + }, + { + "epoch": 0.021214201035752167, + "grad_norm": 5.144644260406494, + "learning_rate": 0.00019858675984276534, + "loss": 0.3185, + "step": 1360 + }, + { + "epoch": 0.021370187808073874, + "grad_norm": 0.3853607177734375, + "learning_rate": 0.00019857636072461056, + "loss": 0.21, + "step": 1370 + }, + { + "epoch": 0.02152617458039558, + "grad_norm": 15.329313278198242, + "learning_rate": 0.00019856596160645578, + "loss": 0.4216, + "step": 1380 + }, + { + "epoch": 0.02168216135271729, + "grad_norm": 1.9162156581878662, + "learning_rate": 0.000198555562488301, + "loss": 0.3766, + "step": 1390 + }, + { + "epoch": 0.021838148125038998, + "grad_norm": 2.5800933837890625, + "learning_rate": 0.00019854516337014622, + "loss": 0.4478, + "step": 1400 + }, + { + "epoch": 0.021994134897360705, + "grad_norm": 12.898444175720215, + "learning_rate": 0.00019853476425199144, + "loss": 0.7923, + "step": 1410 + }, + { + "epoch": 0.022150121669682412, + "grad_norm": 11.576621055603027, + "learning_rate": 0.00019852436513383666, + "loss": 0.3582, + "step": 1420 + }, + { + "epoch": 0.02230610844200412, + "grad_norm": 10.80225658416748, + "learning_rate": 0.00019851396601568188, + "loss": 0.4698, + "step": 1430 + }, + { + "epoch": 0.022462095214325826, + "grad_norm": 5.019324779510498, + "learning_rate": 0.0001985035668975271, + "loss": 0.4463, + "step": 1440 + }, + { + "epoch": 0.022618081986647533, + "grad_norm": 5.366154670715332, + "learning_rate": 0.00019849316777937232, + "loss": 0.4288, + "step": 1450 + }, + { + "epoch": 0.02277406875896924, + "grad_norm": 3.945873498916626, + "learning_rate": 0.00019848276866121754, + "loss": 0.5625, + "step": 1460 + }, + { + "epoch": 0.022930055531290947, + "grad_norm": 1.8938435316085815, + "learning_rate": 0.00019847236954306276, + "loss": 0.4298, + "step": 1470 + }, + { + "epoch": 0.023086042303612653, + "grad_norm": 7.040036678314209, + "learning_rate": 0.00019846197042490798, + "loss": 0.5806, + "step": 1480 + }, + { + "epoch": 0.02324202907593436, + "grad_norm": 12.009136199951172, + "learning_rate": 0.0001984515713067532, + "loss": 0.3603, + "step": 1490 + }, + { + "epoch": 0.023398015848256067, + "grad_norm": 3.157111883163452, + "learning_rate": 0.00019844117218859841, + "loss": 0.3811, + "step": 1500 + }, + { + "epoch": 0.023554002620577774, + "grad_norm": 1.0227491855621338, + "learning_rate": 0.00019843077307044363, + "loss": 0.3459, + "step": 1510 + }, + { + "epoch": 0.02370998939289948, + "grad_norm": 2.699263334274292, + "learning_rate": 0.00019842037395228885, + "loss": 0.4047, + "step": 1520 + }, + { + "epoch": 0.023865976165221188, + "grad_norm": 3.879258871078491, + "learning_rate": 0.00019840997483413407, + "loss": 0.2083, + "step": 1530 + }, + { + "epoch": 0.024021962937542895, + "grad_norm": 1.0468460321426392, + "learning_rate": 0.0001983995757159793, + "loss": 0.3146, + "step": 1540 + }, + { + "epoch": 0.024177949709864602, + "grad_norm": 3.5738143920898438, + "learning_rate": 0.0001983891765978245, + "loss": 0.3181, + "step": 1550 + }, + { + "epoch": 0.024333936482186312, + "grad_norm": 6.3198113441467285, + "learning_rate": 0.00019837877747966973, + "loss": 0.3196, + "step": 1560 + }, + { + "epoch": 0.02448992325450802, + "grad_norm": 1.6974947452545166, + "learning_rate": 0.00019836837836151495, + "loss": 0.3832, + "step": 1570 + }, + { + "epoch": 0.024645910026829726, + "grad_norm": 6.220627307891846, + "learning_rate": 0.00019835797924336017, + "loss": 0.3445, + "step": 1580 + }, + { + "epoch": 0.024801896799151433, + "grad_norm": 3.9202497005462646, + "learning_rate": 0.0001983475801252054, + "loss": 0.4304, + "step": 1590 + }, + { + "epoch": 0.02495788357147314, + "grad_norm": 2.2945311069488525, + "learning_rate": 0.0001983371810070506, + "loss": 0.4487, + "step": 1600 + }, + { + "epoch": 0.025113870343794847, + "grad_norm": 9.262955665588379, + "learning_rate": 0.00019832678188889583, + "loss": 0.3691, + "step": 1610 + }, + { + "epoch": 0.025269857116116554, + "grad_norm": 2.6905221939086914, + "learning_rate": 0.00019831638277074105, + "loss": 0.5117, + "step": 1620 + }, + { + "epoch": 0.02542584388843826, + "grad_norm": 4.1330060958862305, + "learning_rate": 0.00019830598365258627, + "loss": 0.5721, + "step": 1630 + }, + { + "epoch": 0.025581830660759967, + "grad_norm": 1.0681806802749634, + "learning_rate": 0.0001982955845344315, + "loss": 0.4794, + "step": 1640 + }, + { + "epoch": 0.025737817433081674, + "grad_norm": 2.6171841621398926, + "learning_rate": 0.0001982851854162767, + "loss": 0.4596, + "step": 1650 + }, + { + "epoch": 0.02589380420540338, + "grad_norm": 3.3451762199401855, + "learning_rate": 0.00019827478629812193, + "loss": 0.4745, + "step": 1660 + }, + { + "epoch": 0.026049790977725088, + "grad_norm": 4.152349472045898, + "learning_rate": 0.00019826438717996715, + "loss": 0.4452, + "step": 1670 + }, + { + "epoch": 0.026205777750046795, + "grad_norm": 2.7109954357147217, + "learning_rate": 0.00019825398806181237, + "loss": 0.3966, + "step": 1680 + }, + { + "epoch": 0.026361764522368502, + "grad_norm": 5.731235980987549, + "learning_rate": 0.0001982435889436576, + "loss": 0.4423, + "step": 1690 + }, + { + "epoch": 0.02651775129469021, + "grad_norm": 2.370852470397949, + "learning_rate": 0.0001982331898255028, + "loss": 0.5619, + "step": 1700 + }, + { + "epoch": 0.026673738067011916, + "grad_norm": 2.5500876903533936, + "learning_rate": 0.00019822279070734803, + "loss": 0.2658, + "step": 1710 + }, + { + "epoch": 0.026829724839333626, + "grad_norm": 0.20157204568386078, + "learning_rate": 0.00019821239158919325, + "loss": 0.2939, + "step": 1720 + }, + { + "epoch": 0.026985711611655333, + "grad_norm": 2.6305084228515625, + "learning_rate": 0.00019820199247103847, + "loss": 0.5339, + "step": 1730 + }, + { + "epoch": 0.02714169838397704, + "grad_norm": 6.241075038909912, + "learning_rate": 0.00019819159335288369, + "loss": 0.3625, + "step": 1740 + }, + { + "epoch": 0.027297685156298747, + "grad_norm": 5.584851264953613, + "learning_rate": 0.0001981811942347289, + "loss": 0.6158, + "step": 1750 + }, + { + "epoch": 0.027453671928620454, + "grad_norm": 6.756530284881592, + "learning_rate": 0.00019817079511657413, + "loss": 0.3513, + "step": 1760 + }, + { + "epoch": 0.02760965870094216, + "grad_norm": 5.153111457824707, + "learning_rate": 0.00019816039599841935, + "loss": 0.2885, + "step": 1770 + }, + { + "epoch": 0.027765645473263868, + "grad_norm": 6.222873210906982, + "learning_rate": 0.00019814999688026456, + "loss": 0.3291, + "step": 1780 + }, + { + "epoch": 0.027921632245585574, + "grad_norm": 4.744391918182373, + "learning_rate": 0.00019813959776210978, + "loss": 0.5053, + "step": 1790 + }, + { + "epoch": 0.02807761901790728, + "grad_norm": 3.7181131839752197, + "learning_rate": 0.000198129198643955, + "loss": 0.3983, + "step": 1800 + }, + { + "epoch": 0.02823360579022899, + "grad_norm": 1.8923834562301636, + "learning_rate": 0.00019811879952580022, + "loss": 0.4465, + "step": 1810 + }, + { + "epoch": 0.028389592562550695, + "grad_norm": 6.049956321716309, + "learning_rate": 0.00019810840040764544, + "loss": 0.5028, + "step": 1820 + }, + { + "epoch": 0.028545579334872402, + "grad_norm": 8.034428596496582, + "learning_rate": 0.00019809800128949066, + "loss": 0.4799, + "step": 1830 + }, + { + "epoch": 0.02870156610719411, + "grad_norm": 3.916890859603882, + "learning_rate": 0.00019808760217133588, + "loss": 0.3109, + "step": 1840 + }, + { + "epoch": 0.028857552879515816, + "grad_norm": 3.351285457611084, + "learning_rate": 0.0001980772030531811, + "loss": 0.5981, + "step": 1850 + }, + { + "epoch": 0.029013539651837523, + "grad_norm": 4.443714141845703, + "learning_rate": 0.00019806680393502632, + "loss": 0.3243, + "step": 1860 + }, + { + "epoch": 0.02916952642415923, + "grad_norm": 0.6998550295829773, + "learning_rate": 0.00019805640481687154, + "loss": 0.2057, + "step": 1870 + }, + { + "epoch": 0.02932551319648094, + "grad_norm": 12.252562522888184, + "learning_rate": 0.00019804600569871676, + "loss": 0.3806, + "step": 1880 + }, + { + "epoch": 0.029481499968802647, + "grad_norm": 9.03868293762207, + "learning_rate": 0.00019803560658056198, + "loss": 0.3351, + "step": 1890 + }, + { + "epoch": 0.029637486741124354, + "grad_norm": 2.3400044441223145, + "learning_rate": 0.0001980252074624072, + "loss": 0.3103, + "step": 1900 + }, + { + "epoch": 0.02979347351344606, + "grad_norm": 2.399489164352417, + "learning_rate": 0.00019801480834425242, + "loss": 0.4792, + "step": 1910 + }, + { + "epoch": 0.029949460285767768, + "grad_norm": 1.5152381658554077, + "learning_rate": 0.00019800440922609764, + "loss": 0.2873, + "step": 1920 + }, + { + "epoch": 0.030105447058089475, + "grad_norm": 2.7013959884643555, + "learning_rate": 0.00019799401010794286, + "loss": 0.381, + "step": 1930 + }, + { + "epoch": 0.03026143383041118, + "grad_norm": 0.7135261297225952, + "learning_rate": 0.00019798361098978808, + "loss": 0.4001, + "step": 1940 + }, + { + "epoch": 0.03041742060273289, + "grad_norm": 15.61109447479248, + "learning_rate": 0.0001979732118716333, + "loss": 0.474, + "step": 1950 + }, + { + "epoch": 0.030573407375054595, + "grad_norm": 3.236668348312378, + "learning_rate": 0.00019796281275347852, + "loss": 0.2149, + "step": 1960 + }, + { + "epoch": 0.030729394147376302, + "grad_norm": 4.722929954528809, + "learning_rate": 0.00019795241363532374, + "loss": 0.5474, + "step": 1970 + }, + { + "epoch": 0.03088538091969801, + "grad_norm": 6.002335548400879, + "learning_rate": 0.00019794201451716896, + "loss": 0.5196, + "step": 1980 + }, + { + "epoch": 0.031041367692019716, + "grad_norm": 5.210022926330566, + "learning_rate": 0.00019793161539901418, + "loss": 0.4741, + "step": 1990 + }, + { + "epoch": 0.031197354464341423, + "grad_norm": 7.984034538269043, + "learning_rate": 0.0001979212162808594, + "loss": 0.5399, + "step": 2000 + }, + { + "epoch": 0.03135334123666313, + "grad_norm": 4.008991241455078, + "learning_rate": 0.00019791081716270462, + "loss": 0.4788, + "step": 2010 + }, + { + "epoch": 0.03150932800898484, + "grad_norm": 4.395316123962402, + "learning_rate": 0.00019790041804454984, + "loss": 0.3456, + "step": 2020 + }, + { + "epoch": 0.03166531478130655, + "grad_norm": 3.9784998893737793, + "learning_rate": 0.00019789001892639506, + "loss": 0.326, + "step": 2030 + }, + { + "epoch": 0.03182130155362825, + "grad_norm": 5.040256023406982, + "learning_rate": 0.00019787961980824028, + "loss": 0.5264, + "step": 2040 + }, + { + "epoch": 0.03197728832594996, + "grad_norm": 2.686203956604004, + "learning_rate": 0.0001978692206900855, + "loss": 0.5461, + "step": 2050 + }, + { + "epoch": 0.032133275098271664, + "grad_norm": 1.8417892456054688, + "learning_rate": 0.00019785882157193071, + "loss": 0.4736, + "step": 2060 + }, + { + "epoch": 0.032289261870593375, + "grad_norm": 3.623887300491333, + "learning_rate": 0.00019784842245377593, + "loss": 0.3742, + "step": 2070 + }, + { + "epoch": 0.03244524864291508, + "grad_norm": 0.5184723734855652, + "learning_rate": 0.00019783802333562115, + "loss": 0.3941, + "step": 2080 + }, + { + "epoch": 0.03260123541523679, + "grad_norm": 7.751046180725098, + "learning_rate": 0.00019782762421746637, + "loss": 0.4609, + "step": 2090 + }, + { + "epoch": 0.03275722218755849, + "grad_norm": 2.547175168991089, + "learning_rate": 0.0001978172250993116, + "loss": 0.3089, + "step": 2100 + }, + { + "epoch": 0.0329132089598802, + "grad_norm": 1.8453168869018555, + "learning_rate": 0.0001978068259811568, + "loss": 0.4371, + "step": 2110 + }, + { + "epoch": 0.033069195732201906, + "grad_norm": 7.033441543579102, + "learning_rate": 0.00019779642686300203, + "loss": 0.3842, + "step": 2120 + }, + { + "epoch": 0.033225182504523616, + "grad_norm": 2.2642929553985596, + "learning_rate": 0.00019778602774484725, + "loss": 0.2069, + "step": 2130 + }, + { + "epoch": 0.03338116927684533, + "grad_norm": 4.043938636779785, + "learning_rate": 0.00019777562862669244, + "loss": 0.4148, + "step": 2140 + }, + { + "epoch": 0.03353715604916703, + "grad_norm": 4.227573871612549, + "learning_rate": 0.0001977652295085377, + "loss": 0.4785, + "step": 2150 + }, + { + "epoch": 0.03369314282148874, + "grad_norm": 2.0808868408203125, + "learning_rate": 0.00019775483039038288, + "loss": 0.3073, + "step": 2160 + }, + { + "epoch": 0.033849129593810444, + "grad_norm": 4.518038749694824, + "learning_rate": 0.00019774443127222813, + "loss": 0.4197, + "step": 2170 + }, + { + "epoch": 0.034005116366132154, + "grad_norm": 2.3731303215026855, + "learning_rate": 0.00019773403215407332, + "loss": 0.2753, + "step": 2180 + }, + { + "epoch": 0.03416110313845386, + "grad_norm": 1.7613986730575562, + "learning_rate": 0.00019772363303591857, + "loss": 0.3834, + "step": 2190 + }, + { + "epoch": 0.03431708991077557, + "grad_norm": 2.9518485069274902, + "learning_rate": 0.00019771323391776376, + "loss": 0.3544, + "step": 2200 + }, + { + "epoch": 0.03447307668309727, + "grad_norm": 4.424034118652344, + "learning_rate": 0.000197702834799609, + "loss": 0.4046, + "step": 2210 + }, + { + "epoch": 0.03462906345541898, + "grad_norm": 2.307690382003784, + "learning_rate": 0.0001976924356814542, + "loss": 0.1445, + "step": 2220 + }, + { + "epoch": 0.034785050227740685, + "grad_norm": 2.536533832550049, + "learning_rate": 0.00019768203656329945, + "loss": 0.2815, + "step": 2230 + }, + { + "epoch": 0.034941037000062396, + "grad_norm": 2.006728172302246, + "learning_rate": 0.00019767163744514464, + "loss": 0.5314, + "step": 2240 + }, + { + "epoch": 0.0350970237723841, + "grad_norm": 4.729264259338379, + "learning_rate": 0.0001976612383269899, + "loss": 0.4547, + "step": 2250 + }, + { + "epoch": 0.03525301054470581, + "grad_norm": 4.600561141967773, + "learning_rate": 0.00019765083920883508, + "loss": 0.3967, + "step": 2260 + }, + { + "epoch": 0.03540899731702751, + "grad_norm": 0.6419815421104431, + "learning_rate": 0.00019764044009068033, + "loss": 0.336, + "step": 2270 + }, + { + "epoch": 0.03556498408934922, + "grad_norm": 5.567695140838623, + "learning_rate": 0.00019763004097252552, + "loss": 0.5644, + "step": 2280 + }, + { + "epoch": 0.035720970861670934, + "grad_norm": 0.5767773389816284, + "learning_rate": 0.00019761964185437077, + "loss": 0.2643, + "step": 2290 + }, + { + "epoch": 0.03587695763399264, + "grad_norm": 4.91269063949585, + "learning_rate": 0.00019760924273621596, + "loss": 0.4198, + "step": 2300 + }, + { + "epoch": 0.03603294440631435, + "grad_norm": 4.422239780426025, + "learning_rate": 0.0001975988436180612, + "loss": 0.5338, + "step": 2310 + }, + { + "epoch": 0.03618893117863605, + "grad_norm": 1.672713041305542, + "learning_rate": 0.0001975884444999064, + "loss": 0.3214, + "step": 2320 + }, + { + "epoch": 0.03634491795095776, + "grad_norm": 4.033807277679443, + "learning_rate": 0.00019757804538175164, + "loss": 0.3593, + "step": 2330 + }, + { + "epoch": 0.036500904723279465, + "grad_norm": 3.198880910873413, + "learning_rate": 0.00019756764626359684, + "loss": 0.3657, + "step": 2340 + }, + { + "epoch": 0.036656891495601175, + "grad_norm": 2.640443801879883, + "learning_rate": 0.00019755724714544208, + "loss": 0.3243, + "step": 2350 + }, + { + "epoch": 0.03681287826792288, + "grad_norm": 1.562350869178772, + "learning_rate": 0.00019754684802728728, + "loss": 0.3362, + "step": 2360 + }, + { + "epoch": 0.03696886504024459, + "grad_norm": 4.379998683929443, + "learning_rate": 0.00019753644890913252, + "loss": 0.541, + "step": 2370 + }, + { + "epoch": 0.03712485181256629, + "grad_norm": 2.529301881790161, + "learning_rate": 0.00019752604979097774, + "loss": 0.5396, + "step": 2380 + }, + { + "epoch": 0.037280838584888, + "grad_norm": 1.7864131927490234, + "learning_rate": 0.00019751565067282296, + "loss": 0.3165, + "step": 2390 + }, + { + "epoch": 0.037436825357209706, + "grad_norm": 1.0566092729568481, + "learning_rate": 0.00019750525155466818, + "loss": 0.4225, + "step": 2400 + }, + { + "epoch": 0.03759281212953142, + "grad_norm": 2.4530510902404785, + "learning_rate": 0.0001974948524365134, + "loss": 0.4281, + "step": 2410 + }, + { + "epoch": 0.03774879890185312, + "grad_norm": 5.363385200500488, + "learning_rate": 0.00019748445331835862, + "loss": 0.4746, + "step": 2420 + }, + { + "epoch": 0.03790478567417483, + "grad_norm": 6.038578510284424, + "learning_rate": 0.00019747405420020384, + "loss": 0.3182, + "step": 2430 + }, + { + "epoch": 0.038060772446496534, + "grad_norm": 9.37709903717041, + "learning_rate": 0.00019746365508204906, + "loss": 0.3288, + "step": 2440 + }, + { + "epoch": 0.038216759218818244, + "grad_norm": 28.81134033203125, + "learning_rate": 0.00019745325596389428, + "loss": 0.2646, + "step": 2450 + }, + { + "epoch": 0.038372745991139955, + "grad_norm": 0.616172730922699, + "learning_rate": 0.0001974428568457395, + "loss": 0.3848, + "step": 2460 + }, + { + "epoch": 0.03852873276346166, + "grad_norm": 2.9943106174468994, + "learning_rate": 0.00019743245772758472, + "loss": 0.4097, + "step": 2470 + }, + { + "epoch": 0.03868471953578337, + "grad_norm": 4.001022815704346, + "learning_rate": 0.00019742205860942994, + "loss": 0.2359, + "step": 2480 + }, + { + "epoch": 0.03884070630810507, + "grad_norm": 3.587982416152954, + "learning_rate": 0.00019741165949127516, + "loss": 0.5829, + "step": 2490 + }, + { + "epoch": 0.03899669308042678, + "grad_norm": 1.906042218208313, + "learning_rate": 0.00019740126037312038, + "loss": 0.336, + "step": 2500 + }, + { + "epoch": 0.039152679852748486, + "grad_norm": 9.536376953125, + "learning_rate": 0.0001973908612549656, + "loss": 0.3468, + "step": 2510 + }, + { + "epoch": 0.039308666625070196, + "grad_norm": 1.582669734954834, + "learning_rate": 0.00019738046213681082, + "loss": 0.3534, + "step": 2520 + }, + { + "epoch": 0.0394646533973919, + "grad_norm": 2.59112286567688, + "learning_rate": 0.00019737006301865604, + "loss": 0.239, + "step": 2530 + }, + { + "epoch": 0.03962064016971361, + "grad_norm": 2.091029167175293, + "learning_rate": 0.00019735966390050126, + "loss": 0.5045, + "step": 2540 + }, + { + "epoch": 0.03977662694203531, + "grad_norm": 5.665968894958496, + "learning_rate": 0.00019734926478234648, + "loss": 0.3729, + "step": 2550 + }, + { + "epoch": 0.039932613714357024, + "grad_norm": 2.2713849544525146, + "learning_rate": 0.0001973388656641917, + "loss": 0.3598, + "step": 2560 + }, + { + "epoch": 0.04008860048667873, + "grad_norm": 3.3854405879974365, + "learning_rate": 0.00019732846654603692, + "loss": 0.2485, + "step": 2570 + }, + { + "epoch": 0.04024458725900044, + "grad_norm": 7.780918598175049, + "learning_rate": 0.00019731806742788214, + "loss": 0.3557, + "step": 2580 + }, + { + "epoch": 0.04040057403132214, + "grad_norm": 2.7871644496917725, + "learning_rate": 0.00019730766830972733, + "loss": 0.7119, + "step": 2590 + }, + { + "epoch": 0.04055656080364385, + "grad_norm": 2.53657603263855, + "learning_rate": 0.00019729726919157258, + "loss": 0.3575, + "step": 2600 + }, + { + "epoch": 0.040712547575965555, + "grad_norm": 0.3270189166069031, + "learning_rate": 0.00019728687007341777, + "loss": 0.3733, + "step": 2610 + }, + { + "epoch": 0.040868534348287265, + "grad_norm": 1.2280256748199463, + "learning_rate": 0.00019727647095526301, + "loss": 0.458, + "step": 2620 + }, + { + "epoch": 0.041024521120608975, + "grad_norm": 0.7863756418228149, + "learning_rate": 0.0001972660718371082, + "loss": 0.6961, + "step": 2630 + }, + { + "epoch": 0.04118050789293068, + "grad_norm": 3.562969207763672, + "learning_rate": 0.00019725567271895345, + "loss": 0.3823, + "step": 2640 + }, + { + "epoch": 0.04133649466525239, + "grad_norm": 2.5304505825042725, + "learning_rate": 0.00019724527360079865, + "loss": 0.4938, + "step": 2650 + }, + { + "epoch": 0.04149248143757409, + "grad_norm": 1.8894665241241455, + "learning_rate": 0.0001972348744826439, + "loss": 0.4397, + "step": 2660 + }, + { + "epoch": 0.0416484682098958, + "grad_norm": 3.460160970687866, + "learning_rate": 0.00019722447536448909, + "loss": 0.3571, + "step": 2670 + }, + { + "epoch": 0.04180445498221751, + "grad_norm": 4.1218061447143555, + "learning_rate": 0.00019721407624633433, + "loss": 0.2929, + "step": 2680 + }, + { + "epoch": 0.04196044175453922, + "grad_norm": 2.9370386600494385, + "learning_rate": 0.00019720367712817952, + "loss": 0.4142, + "step": 2690 + }, + { + "epoch": 0.04211642852686092, + "grad_norm": 5.4618401527404785, + "learning_rate": 0.00019719327801002477, + "loss": 0.3288, + "step": 2700 + }, + { + "epoch": 0.04227241529918263, + "grad_norm": 3.1975386142730713, + "learning_rate": 0.00019718287889186996, + "loss": 0.3373, + "step": 2710 + }, + { + "epoch": 0.042428402071504334, + "grad_norm": 4.580129623413086, + "learning_rate": 0.0001971724797737152, + "loss": 0.7003, + "step": 2720 + }, + { + "epoch": 0.042584388843826045, + "grad_norm": 1.7264137268066406, + "learning_rate": 0.0001971620806555604, + "loss": 0.3375, + "step": 2730 + }, + { + "epoch": 0.04274037561614775, + "grad_norm": 8.90623950958252, + "learning_rate": 0.00019715168153740565, + "loss": 0.4492, + "step": 2740 + }, + { + "epoch": 0.04289636238846946, + "grad_norm": 8.407225608825684, + "learning_rate": 0.00019714128241925084, + "loss": 0.3827, + "step": 2750 + }, + { + "epoch": 0.04305234916079116, + "grad_norm": 5.202430725097656, + "learning_rate": 0.0001971308833010961, + "loss": 0.2806, + "step": 2760 + }, + { + "epoch": 0.04320833593311287, + "grad_norm": 1.9569600820541382, + "learning_rate": 0.00019712048418294128, + "loss": 0.2506, + "step": 2770 + }, + { + "epoch": 0.04336432270543458, + "grad_norm": 3.2171106338500977, + "learning_rate": 0.00019711008506478653, + "loss": 0.3344, + "step": 2780 + }, + { + "epoch": 0.043520309477756286, + "grad_norm": 0.7588744163513184, + "learning_rate": 0.00019709968594663172, + "loss": 0.3656, + "step": 2790 + }, + { + "epoch": 0.043676296250077996, + "grad_norm": 1.4201767444610596, + "learning_rate": 0.00019708928682847697, + "loss": 0.4033, + "step": 2800 + }, + { + "epoch": 0.0438322830223997, + "grad_norm": 2.439276933670044, + "learning_rate": 0.00019707888771032216, + "loss": 0.2828, + "step": 2810 + }, + { + "epoch": 0.04398826979472141, + "grad_norm": 18.436410903930664, + "learning_rate": 0.0001970684885921674, + "loss": 0.6879, + "step": 2820 + }, + { + "epoch": 0.044144256567043114, + "grad_norm": 3.1035194396972656, + "learning_rate": 0.0001970580894740126, + "loss": 0.3022, + "step": 2830 + }, + { + "epoch": 0.044300243339364824, + "grad_norm": 4.849295139312744, + "learning_rate": 0.00019704769035585785, + "loss": 0.4078, + "step": 2840 + }, + { + "epoch": 0.04445623011168653, + "grad_norm": 3.680732488632202, + "learning_rate": 0.00019703729123770304, + "loss": 0.4738, + "step": 2850 + }, + { + "epoch": 0.04461221688400824, + "grad_norm": 2.8644354343414307, + "learning_rate": 0.00019702689211954829, + "loss": 0.386, + "step": 2860 + }, + { + "epoch": 0.04476820365632994, + "grad_norm": 0.7621586918830872, + "learning_rate": 0.00019701649300139348, + "loss": 0.2032, + "step": 2870 + }, + { + "epoch": 0.04492419042865165, + "grad_norm": 0.9556747078895569, + "learning_rate": 0.00019700609388323873, + "loss": 0.5366, + "step": 2880 + }, + { + "epoch": 0.045080177200973355, + "grad_norm": 4.282146453857422, + "learning_rate": 0.00019699569476508392, + "loss": 0.4814, + "step": 2890 + }, + { + "epoch": 0.045236163973295065, + "grad_norm": 1.905647873878479, + "learning_rate": 0.00019698529564692916, + "loss": 0.5273, + "step": 2900 + }, + { + "epoch": 0.04539215074561677, + "grad_norm": 0.7890440821647644, + "learning_rate": 0.00019697489652877436, + "loss": 0.3434, + "step": 2910 + }, + { + "epoch": 0.04554813751793848, + "grad_norm": 1.717529296875, + "learning_rate": 0.0001969644974106196, + "loss": 0.6094, + "step": 2920 + }, + { + "epoch": 0.04570412429026018, + "grad_norm": 3.9639952182769775, + "learning_rate": 0.0001969540982924648, + "loss": 0.2958, + "step": 2930 + }, + { + "epoch": 0.04586011106258189, + "grad_norm": 1.4309637546539307, + "learning_rate": 0.00019694369917431004, + "loss": 0.2879, + "step": 2940 + }, + { + "epoch": 0.0460160978349036, + "grad_norm": 8.747053146362305, + "learning_rate": 0.00019693330005615524, + "loss": 0.5953, + "step": 2950 + }, + { + "epoch": 0.04617208460722531, + "grad_norm": 0.027920352295041084, + "learning_rate": 0.00019692290093800048, + "loss": 0.2181, + "step": 2960 + }, + { + "epoch": 0.04632807137954702, + "grad_norm": 1.9787837266921997, + "learning_rate": 0.00019691250181984567, + "loss": 0.2764, + "step": 2970 + }, + { + "epoch": 0.04648405815186872, + "grad_norm": 10.669599533081055, + "learning_rate": 0.00019690210270169092, + "loss": 0.5824, + "step": 2980 + }, + { + "epoch": 0.04664004492419043, + "grad_norm": 2.9555306434631348, + "learning_rate": 0.00019689170358353611, + "loss": 0.5993, + "step": 2990 + }, + { + "epoch": 0.046796031696512135, + "grad_norm": 2.168975353240967, + "learning_rate": 0.00019688130446538136, + "loss": 0.2988, + "step": 3000 + }, + { + "epoch": 0.046952018468833845, + "grad_norm": 1.4161510467529297, + "learning_rate": 0.00019687090534722655, + "loss": 0.2984, + "step": 3010 + }, + { + "epoch": 0.04710800524115555, + "grad_norm": 4.282196521759033, + "learning_rate": 0.0001968605062290718, + "loss": 0.2822, + "step": 3020 + }, + { + "epoch": 0.04726399201347726, + "grad_norm": 2.5621113777160645, + "learning_rate": 0.000196850107110917, + "loss": 0.3886, + "step": 3030 + }, + { + "epoch": 0.04741997878579896, + "grad_norm": 0.7100577354431152, + "learning_rate": 0.0001968397079927622, + "loss": 0.2256, + "step": 3040 + }, + { + "epoch": 0.04757596555812067, + "grad_norm": 2.2435333728790283, + "learning_rate": 0.00019682930887460743, + "loss": 0.2155, + "step": 3050 + }, + { + "epoch": 0.047731952330442376, + "grad_norm": 1.9456003904342651, + "learning_rate": 0.00019681890975645265, + "loss": 0.2938, + "step": 3060 + }, + { + "epoch": 0.047887939102764086, + "grad_norm": 8.851899147033691, + "learning_rate": 0.00019680851063829787, + "loss": 0.3822, + "step": 3070 + }, + { + "epoch": 0.04804392587508579, + "grad_norm": 3.1571285724639893, + "learning_rate": 0.0001967981115201431, + "loss": 0.4913, + "step": 3080 + }, + { + "epoch": 0.0481999126474075, + "grad_norm": 6.932368755340576, + "learning_rate": 0.0001967877124019883, + "loss": 0.2643, + "step": 3090 + }, + { + "epoch": 0.048355899419729204, + "grad_norm": 4.671253204345703, + "learning_rate": 0.00019677731328383353, + "loss": 0.3736, + "step": 3100 + }, + { + "epoch": 0.048511886192050914, + "grad_norm": 4.238051891326904, + "learning_rate": 0.00019676691416567875, + "loss": 0.2905, + "step": 3110 + }, + { + "epoch": 0.048667872964372624, + "grad_norm": 4.185165882110596, + "learning_rate": 0.00019675651504752397, + "loss": 0.5799, + "step": 3120 + }, + { + "epoch": 0.04882385973669433, + "grad_norm": 5.000885486602783, + "learning_rate": 0.0001967461159293692, + "loss": 0.5353, + "step": 3130 + }, + { + "epoch": 0.04897984650901604, + "grad_norm": 2.1454551219940186, + "learning_rate": 0.0001967357168112144, + "loss": 0.2612, + "step": 3140 + }, + { + "epoch": 0.04913583328133774, + "grad_norm": 5.727521896362305, + "learning_rate": 0.00019672531769305963, + "loss": 0.3584, + "step": 3150 + }, + { + "epoch": 0.04929182005365945, + "grad_norm": 2.14788556098938, + "learning_rate": 0.00019671491857490485, + "loss": 0.248, + "step": 3160 + }, + { + "epoch": 0.049447806825981155, + "grad_norm": 5.505538463592529, + "learning_rate": 0.00019670451945675007, + "loss": 0.5042, + "step": 3170 + }, + { + "epoch": 0.049603793598302866, + "grad_norm": 0.45676565170288086, + "learning_rate": 0.0001966941203385953, + "loss": 0.3119, + "step": 3180 + }, + { + "epoch": 0.04975978037062457, + "grad_norm": 2.446408748626709, + "learning_rate": 0.0001966837212204405, + "loss": 0.1164, + "step": 3190 + }, + { + "epoch": 0.04991576714294628, + "grad_norm": 0.8342123031616211, + "learning_rate": 0.00019667332210228573, + "loss": 0.3749, + "step": 3200 + }, + { + "epoch": 0.05007175391526798, + "grad_norm": 3.0448403358459473, + "learning_rate": 0.00019666292298413095, + "loss": 0.4866, + "step": 3210 + }, + { + "epoch": 0.05022774068758969, + "grad_norm": 0.2389092594385147, + "learning_rate": 0.00019665252386597617, + "loss": 0.4043, + "step": 3220 + }, + { + "epoch": 0.0503837274599114, + "grad_norm": 5.156858921051025, + "learning_rate": 0.00019664212474782139, + "loss": 0.2329, + "step": 3230 + }, + { + "epoch": 0.05053971423223311, + "grad_norm": 3.149811029434204, + "learning_rate": 0.0001966317256296666, + "loss": 0.3312, + "step": 3240 + }, + { + "epoch": 0.05069570100455481, + "grad_norm": 0.45096033811569214, + "learning_rate": 0.00019662132651151185, + "loss": 0.3754, + "step": 3250 + }, + { + "epoch": 0.05085168777687652, + "grad_norm": 0.986663818359375, + "learning_rate": 0.00019661092739335704, + "loss": 0.3235, + "step": 3260 + }, + { + "epoch": 0.05100767454919823, + "grad_norm": 2.9946937561035156, + "learning_rate": 0.0001966005282752023, + "loss": 0.2073, + "step": 3270 + }, + { + "epoch": 0.051163661321519935, + "grad_norm": 3.995633125305176, + "learning_rate": 0.00019659012915704748, + "loss": 0.3481, + "step": 3280 + }, + { + "epoch": 0.051319648093841645, + "grad_norm": 1.7635934352874756, + "learning_rate": 0.00019657973003889273, + "loss": 0.1994, + "step": 3290 + }, + { + "epoch": 0.05147563486616335, + "grad_norm": 4.386302471160889, + "learning_rate": 0.00019656933092073792, + "loss": 0.3384, + "step": 3300 + }, + { + "epoch": 0.05163162163848506, + "grad_norm": 2.7013816833496094, + "learning_rate": 0.00019655893180258317, + "loss": 0.2823, + "step": 3310 + }, + { + "epoch": 0.05178760841080676, + "grad_norm": 7.045024871826172, + "learning_rate": 0.00019654853268442836, + "loss": 0.2112, + "step": 3320 + }, + { + "epoch": 0.05194359518312847, + "grad_norm": 2.587613105773926, + "learning_rate": 0.0001965381335662736, + "loss": 0.3959, + "step": 3330 + }, + { + "epoch": 0.052099581955450176, + "grad_norm": 0.9898849129676819, + "learning_rate": 0.0001965277344481188, + "loss": 0.3295, + "step": 3340 + }, + { + "epoch": 0.05225556872777189, + "grad_norm": 8.89372444152832, + "learning_rate": 0.00019651733532996405, + "loss": 0.3443, + "step": 3350 + }, + { + "epoch": 0.05241155550009359, + "grad_norm": 3.0839977264404297, + "learning_rate": 0.00019650693621180924, + "loss": 0.4376, + "step": 3360 + }, + { + "epoch": 0.0525675422724153, + "grad_norm": 5.862309455871582, + "learning_rate": 0.0001964965370936545, + "loss": 0.5446, + "step": 3370 + }, + { + "epoch": 0.052723529044737004, + "grad_norm": 5.274170398712158, + "learning_rate": 0.00019648613797549968, + "loss": 0.3987, + "step": 3380 + }, + { + "epoch": 0.052879515817058714, + "grad_norm": 4.794046878814697, + "learning_rate": 0.00019647573885734493, + "loss": 0.5267, + "step": 3390 + }, + { + "epoch": 0.05303550258938042, + "grad_norm": 1.6140331029891968, + "learning_rate": 0.00019646533973919012, + "loss": 0.6051, + "step": 3400 + }, + { + "epoch": 0.05319148936170213, + "grad_norm": 6.158959865570068, + "learning_rate": 0.00019645494062103537, + "loss": 0.3056, + "step": 3410 + }, + { + "epoch": 0.05334747613402383, + "grad_norm": 1.6302224397659302, + "learning_rate": 0.00019644454150288056, + "loss": 0.3473, + "step": 3420 + }, + { + "epoch": 0.05350346290634554, + "grad_norm": 4.881719589233398, + "learning_rate": 0.0001964341423847258, + "loss": 0.4032, + "step": 3430 + }, + { + "epoch": 0.05365944967866725, + "grad_norm": 1.1205174922943115, + "learning_rate": 0.000196423743266571, + "loss": 0.4288, + "step": 3440 + }, + { + "epoch": 0.053815436450988956, + "grad_norm": 6.742442607879639, + "learning_rate": 0.00019641334414841624, + "loss": 0.3299, + "step": 3450 + }, + { + "epoch": 0.053971423223310666, + "grad_norm": 0.43702423572540283, + "learning_rate": 0.00019640294503026144, + "loss": 0.4711, + "step": 3460 + }, + { + "epoch": 0.05412740999563237, + "grad_norm": 0.10137466341257095, + "learning_rate": 0.00019639254591210668, + "loss": 0.3355, + "step": 3470 + }, + { + "epoch": 0.05428339676795408, + "grad_norm": 3.7362818717956543, + "learning_rate": 0.00019638214679395188, + "loss": 0.3051, + "step": 3480 + }, + { + "epoch": 0.05443938354027578, + "grad_norm": 2.639213800430298, + "learning_rate": 0.0001963717476757971, + "loss": 0.3746, + "step": 3490 + }, + { + "epoch": 0.054595370312597494, + "grad_norm": 0.7321776151657104, + "learning_rate": 0.00019636134855764232, + "loss": 0.4572, + "step": 3500 + }, + { + "epoch": 0.0547513570849192, + "grad_norm": 1.803120732307434, + "learning_rate": 0.00019635094943948754, + "loss": 0.2899, + "step": 3510 + }, + { + "epoch": 0.05490734385724091, + "grad_norm": 0.9119880795478821, + "learning_rate": 0.00019634055032133276, + "loss": 0.2634, + "step": 3520 + }, + { + "epoch": 0.05506333062956261, + "grad_norm": 14.491175651550293, + "learning_rate": 0.00019633015120317797, + "loss": 0.4814, + "step": 3530 + }, + { + "epoch": 0.05521931740188432, + "grad_norm": 1.126238465309143, + "learning_rate": 0.0001963197520850232, + "loss": 0.3872, + "step": 3540 + }, + { + "epoch": 0.055375304174206025, + "grad_norm": 1.937894582748413, + "learning_rate": 0.00019630935296686841, + "loss": 0.4281, + "step": 3550 + }, + { + "epoch": 0.055531290946527735, + "grad_norm": 3.4557878971099854, + "learning_rate": 0.00019629895384871363, + "loss": 0.3678, + "step": 3560 + }, + { + "epoch": 0.05568727771884944, + "grad_norm": 0.7918877005577087, + "learning_rate": 0.00019628855473055885, + "loss": 0.1537, + "step": 3570 + }, + { + "epoch": 0.05584326449117115, + "grad_norm": 0.9775216579437256, + "learning_rate": 0.00019627815561240407, + "loss": 0.3411, + "step": 3580 + }, + { + "epoch": 0.05599925126349285, + "grad_norm": 3.5335514545440674, + "learning_rate": 0.0001962677564942493, + "loss": 0.4431, + "step": 3590 + }, + { + "epoch": 0.05615523803581456, + "grad_norm": 5.357288360595703, + "learning_rate": 0.0001962573573760945, + "loss": 0.5486, + "step": 3600 + }, + { + "epoch": 0.05631122480813627, + "grad_norm": 2.541792869567871, + "learning_rate": 0.00019624695825793973, + "loss": 0.2704, + "step": 3610 + }, + { + "epoch": 0.05646721158045798, + "grad_norm": 1.9300107955932617, + "learning_rate": 0.00019623655913978495, + "loss": 0.536, + "step": 3620 + }, + { + "epoch": 0.05662319835277969, + "grad_norm": 1.5524200201034546, + "learning_rate": 0.00019622616002163017, + "loss": 0.4837, + "step": 3630 + }, + { + "epoch": 0.05677918512510139, + "grad_norm": 4.605470180511475, + "learning_rate": 0.0001962157609034754, + "loss": 0.3935, + "step": 3640 + }, + { + "epoch": 0.0569351718974231, + "grad_norm": 1.5231540203094482, + "learning_rate": 0.0001962053617853206, + "loss": 0.289, + "step": 3650 + }, + { + "epoch": 0.057091158669744804, + "grad_norm": 1.3818199634552002, + "learning_rate": 0.00019619496266716583, + "loss": 0.3829, + "step": 3660 + }, + { + "epoch": 0.057247145442066515, + "grad_norm": 2.465237617492676, + "learning_rate": 0.00019618456354901105, + "loss": 0.4775, + "step": 3670 + }, + { + "epoch": 0.05740313221438822, + "grad_norm": 2.287116765975952, + "learning_rate": 0.00019617416443085627, + "loss": 0.6419, + "step": 3680 + }, + { + "epoch": 0.05755911898670993, + "grad_norm": 0.8104041218757629, + "learning_rate": 0.0001961637653127015, + "loss": 0.4125, + "step": 3690 + }, + { + "epoch": 0.05771510575903163, + "grad_norm": 6.393253803253174, + "learning_rate": 0.0001961533661945467, + "loss": 0.3983, + "step": 3700 + }, + { + "epoch": 0.05787109253135334, + "grad_norm": 1.9173061847686768, + "learning_rate": 0.00019614296707639193, + "loss": 0.3925, + "step": 3710 + }, + { + "epoch": 0.058027079303675046, + "grad_norm": 8.817895889282227, + "learning_rate": 0.00019613256795823715, + "loss": 0.6599, + "step": 3720 + }, + { + "epoch": 0.058183066075996756, + "grad_norm": 1.728727102279663, + "learning_rate": 0.00019612216884008237, + "loss": 0.4401, + "step": 3730 + }, + { + "epoch": 0.05833905284831846, + "grad_norm": 4.851232051849365, + "learning_rate": 0.0001961117697219276, + "loss": 0.1871, + "step": 3740 + }, + { + "epoch": 0.05849503962064017, + "grad_norm": 4.049133777618408, + "learning_rate": 0.0001961013706037728, + "loss": 0.3421, + "step": 3750 + }, + { + "epoch": 0.05865102639296188, + "grad_norm": 4.8147149085998535, + "learning_rate": 0.00019609097148561803, + "loss": 0.3399, + "step": 3760 + }, + { + "epoch": 0.058807013165283584, + "grad_norm": 2.2623703479766846, + "learning_rate": 0.00019608057236746325, + "loss": 0.3337, + "step": 3770 + }, + { + "epoch": 0.058962999937605294, + "grad_norm": 3.290314197540283, + "learning_rate": 0.00019607017324930847, + "loss": 0.3188, + "step": 3780 + }, + { + "epoch": 0.059118986709927, + "grad_norm": 6.715724945068359, + "learning_rate": 0.00019605977413115369, + "loss": 0.4027, + "step": 3790 + }, + { + "epoch": 0.05927497348224871, + "grad_norm": 0.1159973219037056, + "learning_rate": 0.0001960493750129989, + "loss": 0.3586, + "step": 3800 + }, + { + "epoch": 0.05943096025457041, + "grad_norm": 6.8083882331848145, + "learning_rate": 0.00019603897589484412, + "loss": 0.436, + "step": 3810 + }, + { + "epoch": 0.05958694702689212, + "grad_norm": 2.4548165798187256, + "learning_rate": 0.00019602857677668934, + "loss": 0.3744, + "step": 3820 + }, + { + "epoch": 0.059742933799213825, + "grad_norm": 7.0474629402160645, + "learning_rate": 0.00019601817765853456, + "loss": 0.4615, + "step": 3830 + }, + { + "epoch": 0.059898920571535536, + "grad_norm": 9.740936279296875, + "learning_rate": 0.00019600777854037978, + "loss": 0.3168, + "step": 3840 + }, + { + "epoch": 0.06005490734385724, + "grad_norm": 1.5505999326705933, + "learning_rate": 0.000195997379422225, + "loss": 0.2565, + "step": 3850 + }, + { + "epoch": 0.06021089411617895, + "grad_norm": 1.97008216381073, + "learning_rate": 0.00019598698030407022, + "loss": 0.4208, + "step": 3860 + }, + { + "epoch": 0.06036688088850065, + "grad_norm": 1.0527325868606567, + "learning_rate": 0.00019597658118591544, + "loss": 0.2885, + "step": 3870 + }, + { + "epoch": 0.06052286766082236, + "grad_norm": 3.9389073848724365, + "learning_rate": 0.00019596618206776066, + "loss": 0.3585, + "step": 3880 + }, + { + "epoch": 0.06067885443314407, + "grad_norm": 1.5215762853622437, + "learning_rate": 0.00019595578294960588, + "loss": 0.2961, + "step": 3890 + }, + { + "epoch": 0.06083484120546578, + "grad_norm": 3.200901985168457, + "learning_rate": 0.0001959453838314511, + "loss": 0.2842, + "step": 3900 + }, + { + "epoch": 0.06099082797778748, + "grad_norm": 3.8170034885406494, + "learning_rate": 0.00019593498471329632, + "loss": 0.3067, + "step": 3910 + }, + { + "epoch": 0.06114681475010919, + "grad_norm": 6.250919342041016, + "learning_rate": 0.00019592458559514154, + "loss": 0.3787, + "step": 3920 + }, + { + "epoch": 0.0613028015224309, + "grad_norm": 0.3816400468349457, + "learning_rate": 0.00019591418647698676, + "loss": 0.3611, + "step": 3930 + }, + { + "epoch": 0.061458788294752605, + "grad_norm": 0.548915445804596, + "learning_rate": 0.00019590378735883198, + "loss": 0.3225, + "step": 3940 + }, + { + "epoch": 0.061614775067074315, + "grad_norm": 3.9601895809173584, + "learning_rate": 0.0001958933882406772, + "loss": 0.3119, + "step": 3950 + }, + { + "epoch": 0.06177076183939602, + "grad_norm": 3.092454195022583, + "learning_rate": 0.00019588298912252242, + "loss": 0.2339, + "step": 3960 + }, + { + "epoch": 0.06192674861171773, + "grad_norm": 3.365994691848755, + "learning_rate": 0.00019587259000436764, + "loss": 0.3607, + "step": 3970 + }, + { + "epoch": 0.06208273538403943, + "grad_norm": 3.847018241882324, + "learning_rate": 0.00019586219088621286, + "loss": 0.2026, + "step": 3980 + }, + { + "epoch": 0.06223872215636114, + "grad_norm": 2.2113258838653564, + "learning_rate": 0.00019585179176805808, + "loss": 0.3062, + "step": 3990 + }, + { + "epoch": 0.062394708928682846, + "grad_norm": 2.4622628688812256, + "learning_rate": 0.0001958413926499033, + "loss": 0.1639, + "step": 4000 + }, + { + "epoch": 0.06255069570100455, + "grad_norm": 1.026604413986206, + "learning_rate": 0.00019583099353174852, + "loss": 0.3698, + "step": 4010 + }, + { + "epoch": 0.06270668247332627, + "grad_norm": 2.184748649597168, + "learning_rate": 0.00019582059441359374, + "loss": 0.4276, + "step": 4020 + }, + { + "epoch": 0.06286266924564797, + "grad_norm": 0.38206374645233154, + "learning_rate": 0.00019581019529543896, + "loss": 0.1884, + "step": 4030 + }, + { + "epoch": 0.06301865601796967, + "grad_norm": 3.477356195449829, + "learning_rate": 0.00019579979617728418, + "loss": 0.4619, + "step": 4040 + }, + { + "epoch": 0.06317464279029138, + "grad_norm": 1.1069543361663818, + "learning_rate": 0.0001957893970591294, + "loss": 0.5837, + "step": 4050 + }, + { + "epoch": 0.0633306295626131, + "grad_norm": 1.015366554260254, + "learning_rate": 0.00019577899794097462, + "loss": 0.2466, + "step": 4060 + }, + { + "epoch": 0.0634866163349348, + "grad_norm": 3.1485137939453125, + "learning_rate": 0.00019576859882281984, + "loss": 0.3343, + "step": 4070 + }, + { + "epoch": 0.0636426031072565, + "grad_norm": 1.8550643920898438, + "learning_rate": 0.00019575819970466505, + "loss": 0.2985, + "step": 4080 + }, + { + "epoch": 0.06379858987957822, + "grad_norm": 5.412409782409668, + "learning_rate": 0.00019574780058651027, + "loss": 0.212, + "step": 4090 + }, + { + "epoch": 0.06395457665189992, + "grad_norm": 6.8558526039123535, + "learning_rate": 0.0001957374014683555, + "loss": 0.7002, + "step": 4100 + }, + { + "epoch": 0.06411056342422163, + "grad_norm": 2.9232373237609863, + "learning_rate": 0.00019572700235020071, + "loss": 0.4317, + "step": 4110 + }, + { + "epoch": 0.06426655019654333, + "grad_norm": 2.8058922290802, + "learning_rate": 0.00019571660323204593, + "loss": 0.3056, + "step": 4120 + }, + { + "epoch": 0.06442253696886505, + "grad_norm": 1.4379881620407104, + "learning_rate": 0.00019570620411389115, + "loss": 0.2719, + "step": 4130 + }, + { + "epoch": 0.06457852374118675, + "grad_norm": 6.666982173919678, + "learning_rate": 0.00019569580499573637, + "loss": 0.2342, + "step": 4140 + }, + { + "epoch": 0.06473451051350845, + "grad_norm": 1.3467093706130981, + "learning_rate": 0.0001956854058775816, + "loss": 0.276, + "step": 4150 + }, + { + "epoch": 0.06489049728583016, + "grad_norm": 2.596644401550293, + "learning_rate": 0.0001956750067594268, + "loss": 0.4239, + "step": 4160 + }, + { + "epoch": 0.06504648405815187, + "grad_norm": 4.598370552062988, + "learning_rate": 0.00019566460764127203, + "loss": 0.4242, + "step": 4170 + }, + { + "epoch": 0.06520247083047358, + "grad_norm": 5.2616424560546875, + "learning_rate": 0.00019565420852311725, + "loss": 0.2431, + "step": 4180 + }, + { + "epoch": 0.06535845760279528, + "grad_norm": 1.9798760414123535, + "learning_rate": 0.00019564380940496247, + "loss": 0.6246, + "step": 4190 + }, + { + "epoch": 0.06551444437511698, + "grad_norm": 1.605912685394287, + "learning_rate": 0.0001956334102868077, + "loss": 0.4617, + "step": 4200 + }, + { + "epoch": 0.0656704311474387, + "grad_norm": 3.7138144969940186, + "learning_rate": 0.0001956230111686529, + "loss": 0.5119, + "step": 4210 + }, + { + "epoch": 0.0658264179197604, + "grad_norm": 2.1792988777160645, + "learning_rate": 0.00019561261205049813, + "loss": 0.4112, + "step": 4220 + }, + { + "epoch": 0.06598240469208211, + "grad_norm": 2.197249174118042, + "learning_rate": 0.00019560221293234335, + "loss": 0.4402, + "step": 4230 + }, + { + "epoch": 0.06613839146440381, + "grad_norm": 2.6887547969818115, + "learning_rate": 0.00019559181381418857, + "loss": 0.2802, + "step": 4240 + }, + { + "epoch": 0.06629437823672553, + "grad_norm": 6.97922945022583, + "learning_rate": 0.0001955814146960338, + "loss": 0.3786, + "step": 4250 + }, + { + "epoch": 0.06645036500904723, + "grad_norm": 2.674954414367676, + "learning_rate": 0.000195571015577879, + "loss": 0.3124, + "step": 4260 + }, + { + "epoch": 0.06660635178136894, + "grad_norm": 1.5316659212112427, + "learning_rate": 0.00019556061645972423, + "loss": 0.2674, + "step": 4270 + }, + { + "epoch": 0.06676233855369065, + "grad_norm": 3.402186393737793, + "learning_rate": 0.00019555021734156945, + "loss": 0.1722, + "step": 4280 + }, + { + "epoch": 0.06691832532601236, + "grad_norm": 4.810988903045654, + "learning_rate": 0.00019553981822341467, + "loss": 0.4486, + "step": 4290 + }, + { + "epoch": 0.06707431209833406, + "grad_norm": 4.834427833557129, + "learning_rate": 0.0001955294191052599, + "loss": 0.3412, + "step": 4300 + }, + { + "epoch": 0.06723029887065576, + "grad_norm": 1.7941811084747314, + "learning_rate": 0.0001955190199871051, + "loss": 0.2732, + "step": 4310 + }, + { + "epoch": 0.06738628564297748, + "grad_norm": 4.941607475280762, + "learning_rate": 0.00019550862086895033, + "loss": 0.3954, + "step": 4320 + }, + { + "epoch": 0.06754227241529918, + "grad_norm": 1.0881506204605103, + "learning_rate": 0.00019549822175079555, + "loss": 0.3729, + "step": 4330 + }, + { + "epoch": 0.06769825918762089, + "grad_norm": 2.2694144248962402, + "learning_rate": 0.00019548782263264077, + "loss": 0.4538, + "step": 4340 + }, + { + "epoch": 0.06785424595994259, + "grad_norm": 2.897859573364258, + "learning_rate": 0.00019547742351448599, + "loss": 0.3101, + "step": 4350 + }, + { + "epoch": 0.06801023273226431, + "grad_norm": 2.17456316947937, + "learning_rate": 0.0001954670243963312, + "loss": 0.5781, + "step": 4360 + }, + { + "epoch": 0.06816621950458601, + "grad_norm": 5.235560417175293, + "learning_rate": 0.00019545662527817642, + "loss": 0.3683, + "step": 4370 + }, + { + "epoch": 0.06832220627690772, + "grad_norm": 3.0128090381622314, + "learning_rate": 0.00019544622616002164, + "loss": 0.2558, + "step": 4380 + }, + { + "epoch": 0.06847819304922942, + "grad_norm": 1.6238577365875244, + "learning_rate": 0.00019543582704186684, + "loss": 0.3038, + "step": 4390 + }, + { + "epoch": 0.06863417982155114, + "grad_norm": 2.9384729862213135, + "learning_rate": 0.00019542542792371208, + "loss": 0.3866, + "step": 4400 + }, + { + "epoch": 0.06879016659387284, + "grad_norm": 0.7549554109573364, + "learning_rate": 0.00019541502880555728, + "loss": 0.2905, + "step": 4410 + }, + { + "epoch": 0.06894615336619454, + "grad_norm": 3.4662294387817383, + "learning_rate": 0.00019540462968740252, + "loss": 0.257, + "step": 4420 + }, + { + "epoch": 0.06910214013851626, + "grad_norm": 4.433900356292725, + "learning_rate": 0.00019539423056924772, + "loss": 0.4666, + "step": 4430 + }, + { + "epoch": 0.06925812691083796, + "grad_norm": 2.6317667961120605, + "learning_rate": 0.00019538383145109296, + "loss": 0.2396, + "step": 4440 + }, + { + "epoch": 0.06941411368315967, + "grad_norm": 0.9813170433044434, + "learning_rate": 0.00019537343233293815, + "loss": 0.2148, + "step": 4450 + }, + { + "epoch": 0.06957010045548137, + "grad_norm": 2.8227760791778564, + "learning_rate": 0.0001953630332147834, + "loss": 0.6245, + "step": 4460 + }, + { + "epoch": 0.06972608722780309, + "grad_norm": 1.489005446434021, + "learning_rate": 0.0001953526340966286, + "loss": 0.5559, + "step": 4470 + }, + { + "epoch": 0.06988207400012479, + "grad_norm": 1.7438796758651733, + "learning_rate": 0.00019534223497847384, + "loss": 0.2688, + "step": 4480 + }, + { + "epoch": 0.0700380607724465, + "grad_norm": 1.6185518503189087, + "learning_rate": 0.00019533183586031903, + "loss": 0.386, + "step": 4490 + }, + { + "epoch": 0.0701940475447682, + "grad_norm": 2.74615740776062, + "learning_rate": 0.00019532143674216428, + "loss": 0.4799, + "step": 4500 + }, + { + "epoch": 0.07035003431708992, + "grad_norm": 4.399687767028809, + "learning_rate": 0.00019531103762400947, + "loss": 0.4494, + "step": 4510 + }, + { + "epoch": 0.07050602108941162, + "grad_norm": 2.602672815322876, + "learning_rate": 0.00019530063850585472, + "loss": 0.2798, + "step": 4520 + }, + { + "epoch": 0.07066200786173332, + "grad_norm": 8.301713943481445, + "learning_rate": 0.0001952902393876999, + "loss": 0.3375, + "step": 4530 + }, + { + "epoch": 0.07081799463405503, + "grad_norm": 4.576874732971191, + "learning_rate": 0.00019527984026954516, + "loss": 0.3633, + "step": 4540 + }, + { + "epoch": 0.07097398140637674, + "grad_norm": 3.145695447921753, + "learning_rate": 0.00019526944115139035, + "loss": 0.4248, + "step": 4550 + }, + { + "epoch": 0.07112996817869845, + "grad_norm": 2.894092082977295, + "learning_rate": 0.0001952590420332356, + "loss": 0.3215, + "step": 4560 + }, + { + "epoch": 0.07128595495102015, + "grad_norm": 0.13097500801086426, + "learning_rate": 0.00019524864291508082, + "loss": 0.3269, + "step": 4570 + }, + { + "epoch": 0.07144194172334187, + "grad_norm": 5.0817341804504395, + "learning_rate": 0.00019523824379692604, + "loss": 0.4399, + "step": 4580 + }, + { + "epoch": 0.07159792849566357, + "grad_norm": 1.6493229866027832, + "learning_rate": 0.00019522784467877126, + "loss": 0.3025, + "step": 4590 + }, + { + "epoch": 0.07175391526798527, + "grad_norm": 6.487491130828857, + "learning_rate": 0.00019521744556061648, + "loss": 0.5133, + "step": 4600 + }, + { + "epoch": 0.07190990204030698, + "grad_norm": 4.299659252166748, + "learning_rate": 0.0001952070464424617, + "loss": 1.2018, + "step": 4610 + }, + { + "epoch": 0.0720658888126287, + "grad_norm": 0.8001397252082825, + "learning_rate": 0.00019519664732430692, + "loss": 0.3898, + "step": 4620 + }, + { + "epoch": 0.0722218755849504, + "grad_norm": 3.345963716506958, + "learning_rate": 0.00019518624820615214, + "loss": 0.3375, + "step": 4630 + }, + { + "epoch": 0.0723778623572721, + "grad_norm": 1.9037283658981323, + "learning_rate": 0.00019517584908799735, + "loss": 0.4469, + "step": 4640 + }, + { + "epoch": 0.0725338491295938, + "grad_norm": 0.64785236120224, + "learning_rate": 0.00019516544996984257, + "loss": 0.2919, + "step": 4650 + }, + { + "epoch": 0.07268983590191552, + "grad_norm": 4.346297740936279, + "learning_rate": 0.0001951550508516878, + "loss": 0.2624, + "step": 4660 + }, + { + "epoch": 0.07284582267423723, + "grad_norm": 1.2748740911483765, + "learning_rate": 0.00019514465173353301, + "loss": 0.2201, + "step": 4670 + }, + { + "epoch": 0.07300180944655893, + "grad_norm": 2.5817744731903076, + "learning_rate": 0.00019513425261537823, + "loss": 0.3252, + "step": 4680 + }, + { + "epoch": 0.07315779621888063, + "grad_norm": 1.5413234233856201, + "learning_rate": 0.00019512385349722345, + "loss": 0.4234, + "step": 4690 + }, + { + "epoch": 0.07331378299120235, + "grad_norm": 1.6661428213119507, + "learning_rate": 0.00019511345437906867, + "loss": 0.5214, + "step": 4700 + }, + { + "epoch": 0.07346976976352405, + "grad_norm": 4.948998928070068, + "learning_rate": 0.0001951030552609139, + "loss": 0.5428, + "step": 4710 + }, + { + "epoch": 0.07362575653584576, + "grad_norm": 1.4078203439712524, + "learning_rate": 0.0001950926561427591, + "loss": 0.4, + "step": 4720 + }, + { + "epoch": 0.07378174330816746, + "grad_norm": 0.6059741973876953, + "learning_rate": 0.00019508225702460433, + "loss": 0.5198, + "step": 4730 + }, + { + "epoch": 0.07393773008048918, + "grad_norm": 0.24869346618652344, + "learning_rate": 0.00019507185790644955, + "loss": 0.3058, + "step": 4740 + }, + { + "epoch": 0.07409371685281088, + "grad_norm": 1.6327784061431885, + "learning_rate": 0.00019506145878829477, + "loss": 1.5225, + "step": 4750 + }, + { + "epoch": 0.07424970362513258, + "grad_norm": 3.5642881393432617, + "learning_rate": 0.00019505105967014, + "loss": 0.387, + "step": 4760 + }, + { + "epoch": 0.0744056903974543, + "grad_norm": 3.4709348678588867, + "learning_rate": 0.0001950406605519852, + "loss": 0.4065, + "step": 4770 + }, + { + "epoch": 0.074561677169776, + "grad_norm": 5.526008129119873, + "learning_rate": 0.00019503026143383043, + "loss": 0.4199, + "step": 4780 + }, + { + "epoch": 0.07471766394209771, + "grad_norm": 7.353578567504883, + "learning_rate": 0.00019501986231567565, + "loss": 0.2635, + "step": 4790 + }, + { + "epoch": 0.07487365071441941, + "grad_norm": 4.601809978485107, + "learning_rate": 0.00019500946319752087, + "loss": 0.4241, + "step": 4800 + }, + { + "epoch": 0.07502963748674113, + "grad_norm": 5.431839466094971, + "learning_rate": 0.0001949990640793661, + "loss": 0.2799, + "step": 4810 + }, + { + "epoch": 0.07518562425906283, + "grad_norm": 3.1551904678344727, + "learning_rate": 0.0001949886649612113, + "loss": 0.4654, + "step": 4820 + }, + { + "epoch": 0.07534161103138454, + "grad_norm": 1.8644407987594604, + "learning_rate": 0.00019497826584305653, + "loss": 0.3813, + "step": 4830 + }, + { + "epoch": 0.07549759780370624, + "grad_norm": 1.2464070320129395, + "learning_rate": 0.00019496786672490175, + "loss": 0.3293, + "step": 4840 + }, + { + "epoch": 0.07565358457602796, + "grad_norm": 3.8643407821655273, + "learning_rate": 0.00019495746760674697, + "loss": 0.5776, + "step": 4850 + }, + { + "epoch": 0.07580957134834966, + "grad_norm": 0.7169101238250732, + "learning_rate": 0.00019494706848859216, + "loss": 0.3103, + "step": 4860 + }, + { + "epoch": 0.07596555812067136, + "grad_norm": 0.9111189246177673, + "learning_rate": 0.0001949366693704374, + "loss": 0.5543, + "step": 4870 + }, + { + "epoch": 0.07612154489299307, + "grad_norm": 5.50398063659668, + "learning_rate": 0.0001949262702522826, + "loss": 0.5258, + "step": 4880 + }, + { + "epoch": 0.07627753166531479, + "grad_norm": 4.419070243835449, + "learning_rate": 0.00019491587113412785, + "loss": 0.6176, + "step": 4890 + }, + { + "epoch": 0.07643351843763649, + "grad_norm": 1.6785311698913574, + "learning_rate": 0.00019490547201597304, + "loss": 0.2027, + "step": 4900 + }, + { + "epoch": 0.07658950520995819, + "grad_norm": 2.111905097961426, + "learning_rate": 0.00019489507289781829, + "loss": 0.193, + "step": 4910 + }, + { + "epoch": 0.07674549198227991, + "grad_norm": 2.166006326675415, + "learning_rate": 0.00019488467377966348, + "loss": 0.1709, + "step": 4920 + }, + { + "epoch": 0.07690147875460161, + "grad_norm": 3.8289716243743896, + "learning_rate": 0.00019487427466150872, + "loss": 0.1576, + "step": 4930 + }, + { + "epoch": 0.07705746552692332, + "grad_norm": 2.0966336727142334, + "learning_rate": 0.00019486387554335392, + "loss": 0.628, + "step": 4940 + }, + { + "epoch": 0.07721345229924502, + "grad_norm": 3.5698137283325195, + "learning_rate": 0.00019485347642519916, + "loss": 0.3256, + "step": 4950 + }, + { + "epoch": 0.07736943907156674, + "grad_norm": 2.1274750232696533, + "learning_rate": 0.00019484307730704436, + "loss": 0.3941, + "step": 4960 + }, + { + "epoch": 0.07752542584388844, + "grad_norm": 4.426887512207031, + "learning_rate": 0.0001948326781888896, + "loss": 0.4366, + "step": 4970 + }, + { + "epoch": 0.07768141261621014, + "grad_norm": 5.6438307762146, + "learning_rate": 0.0001948222790707348, + "loss": 0.2416, + "step": 4980 + }, + { + "epoch": 0.07783739938853185, + "grad_norm": 4.310532093048096, + "learning_rate": 0.00019481187995258004, + "loss": 0.4948, + "step": 4990 + }, + { + "epoch": 0.07799338616085356, + "grad_norm": 1.9643551111221313, + "learning_rate": 0.00019480148083442523, + "loss": 0.2636, + "step": 5000 + }, + { + "epoch": 0.07814937293317527, + "grad_norm": 2.5053608417510986, + "learning_rate": 0.00019479108171627048, + "loss": 0.2981, + "step": 5010 + }, + { + "epoch": 0.07830535970549697, + "grad_norm": 1.290313482284546, + "learning_rate": 0.00019478068259811567, + "loss": 0.2891, + "step": 5020 + }, + { + "epoch": 0.07846134647781867, + "grad_norm": 0.16381753981113434, + "learning_rate": 0.00019477028347996092, + "loss": 0.3624, + "step": 5030 + }, + { + "epoch": 0.07861733325014039, + "grad_norm": 3.1533336639404297, + "learning_rate": 0.0001947598843618061, + "loss": 0.2191, + "step": 5040 + }, + { + "epoch": 0.0787733200224621, + "grad_norm": 0.9760684370994568, + "learning_rate": 0.00019474948524365136, + "loss": 0.464, + "step": 5050 + }, + { + "epoch": 0.0789293067947838, + "grad_norm": 2.2214372158050537, + "learning_rate": 0.00019473908612549655, + "loss": 0.4851, + "step": 5060 + }, + { + "epoch": 0.07908529356710552, + "grad_norm": 1.9488602876663208, + "learning_rate": 0.0001947286870073418, + "loss": 0.3167, + "step": 5070 + }, + { + "epoch": 0.07924128033942722, + "grad_norm": 3.411989212036133, + "learning_rate": 0.000194718287889187, + "loss": 0.3631, + "step": 5080 + }, + { + "epoch": 0.07939726711174892, + "grad_norm": 3.581641912460327, + "learning_rate": 0.00019470788877103224, + "loss": 0.1301, + "step": 5090 + }, + { + "epoch": 0.07955325388407063, + "grad_norm": 3.8233249187469482, + "learning_rate": 0.00019469748965287743, + "loss": 0.277, + "step": 5100 + }, + { + "epoch": 0.07970924065639234, + "grad_norm": 8.081124305725098, + "learning_rate": 0.00019468709053472268, + "loss": 0.2866, + "step": 5110 + }, + { + "epoch": 0.07986522742871405, + "grad_norm": 3.1644890308380127, + "learning_rate": 0.00019467669141656787, + "loss": 0.3535, + "step": 5120 + }, + { + "epoch": 0.08002121420103575, + "grad_norm": 0.956291913986206, + "learning_rate": 0.00019466629229841312, + "loss": 0.4388, + "step": 5130 + }, + { + "epoch": 0.08017720097335745, + "grad_norm": 2.458174467086792, + "learning_rate": 0.0001946558931802583, + "loss": 0.321, + "step": 5140 + }, + { + "epoch": 0.08033318774567917, + "grad_norm": 1.813386082649231, + "learning_rate": 0.00019464549406210356, + "loss": 0.4277, + "step": 5150 + }, + { + "epoch": 0.08048917451800087, + "grad_norm": 0.2468508780002594, + "learning_rate": 0.00019463509494394875, + "loss": 0.3927, + "step": 5160 + }, + { + "epoch": 0.08064516129032258, + "grad_norm": 0.5065591931343079, + "learning_rate": 0.000194624695825794, + "loss": 0.175, + "step": 5170 + }, + { + "epoch": 0.08080114806264428, + "grad_norm": 3.965261697769165, + "learning_rate": 0.0001946142967076392, + "loss": 0.2932, + "step": 5180 + }, + { + "epoch": 0.080957134834966, + "grad_norm": 8.131027221679688, + "learning_rate": 0.00019460389758948444, + "loss": 0.3341, + "step": 5190 + }, + { + "epoch": 0.0811131216072877, + "grad_norm": 3.0608482360839844, + "learning_rate": 0.00019459349847132963, + "loss": 0.337, + "step": 5200 + }, + { + "epoch": 0.0812691083796094, + "grad_norm": 1.2423006296157837, + "learning_rate": 0.00019458309935317487, + "loss": 0.439, + "step": 5210 + }, + { + "epoch": 0.08142509515193111, + "grad_norm": 2.6871418952941895, + "learning_rate": 0.00019457270023502007, + "loss": 0.3247, + "step": 5220 + }, + { + "epoch": 0.08158108192425283, + "grad_norm": 3.4826762676239014, + "learning_rate": 0.00019456230111686531, + "loss": 0.5614, + "step": 5230 + }, + { + "epoch": 0.08173706869657453, + "grad_norm": 1.9816076755523682, + "learning_rate": 0.0001945519019987105, + "loss": 0.5613, + "step": 5240 + }, + { + "epoch": 0.08189305546889623, + "grad_norm": 5.499810218811035, + "learning_rate": 0.00019454150288055575, + "loss": 0.4791, + "step": 5250 + }, + { + "epoch": 0.08204904224121795, + "grad_norm": 3.7328336238861084, + "learning_rate": 0.00019453110376240095, + "loss": 0.2752, + "step": 5260 + }, + { + "epoch": 0.08220502901353965, + "grad_norm": 1.268531084060669, + "learning_rate": 0.0001945207046442462, + "loss": 0.2227, + "step": 5270 + }, + { + "epoch": 0.08236101578586136, + "grad_norm": 4.293695449829102, + "learning_rate": 0.00019451030552609138, + "loss": 0.6035, + "step": 5280 + }, + { + "epoch": 0.08251700255818306, + "grad_norm": 3.009061574935913, + "learning_rate": 0.00019449990640793663, + "loss": 0.4032, + "step": 5290 + }, + { + "epoch": 0.08267298933050478, + "grad_norm": 2.217053174972534, + "learning_rate": 0.00019448950728978182, + "loss": 0.4399, + "step": 5300 + }, + { + "epoch": 0.08282897610282648, + "grad_norm": 2.3602845668792725, + "learning_rate": 0.00019447910817162704, + "loss": 0.6428, + "step": 5310 + }, + { + "epoch": 0.08298496287514819, + "grad_norm": 11.570981979370117, + "learning_rate": 0.00019446870905347226, + "loss": 0.4949, + "step": 5320 + }, + { + "epoch": 0.08314094964746989, + "grad_norm": 4.000115394592285, + "learning_rate": 0.00019445830993531748, + "loss": 0.7965, + "step": 5330 + }, + { + "epoch": 0.0832969364197916, + "grad_norm": 0.9110697507858276, + "learning_rate": 0.0001944479108171627, + "loss": 0.2803, + "step": 5340 + }, + { + "epoch": 0.08345292319211331, + "grad_norm": 2.672752618789673, + "learning_rate": 0.00019443751169900792, + "loss": 0.2366, + "step": 5350 + }, + { + "epoch": 0.08360890996443501, + "grad_norm": 4.964208126068115, + "learning_rate": 0.00019442711258085314, + "loss": 0.3474, + "step": 5360 + }, + { + "epoch": 0.08376489673675672, + "grad_norm": 1.1511766910552979, + "learning_rate": 0.00019441671346269836, + "loss": 0.2477, + "step": 5370 + }, + { + "epoch": 0.08392088350907843, + "grad_norm": 4.001786708831787, + "learning_rate": 0.00019440631434454358, + "loss": 0.349, + "step": 5380 + }, + { + "epoch": 0.08407687028140014, + "grad_norm": 1.4430139064788818, + "learning_rate": 0.0001943959152263888, + "loss": 0.4473, + "step": 5390 + }, + { + "epoch": 0.08423285705372184, + "grad_norm": 2.817662239074707, + "learning_rate": 0.00019438551610823402, + "loss": 1.5236, + "step": 5400 + }, + { + "epoch": 0.08438884382604356, + "grad_norm": 1.8846608400344849, + "learning_rate": 0.00019437511699007924, + "loss": 0.3512, + "step": 5410 + }, + { + "epoch": 0.08454483059836526, + "grad_norm": 0.6527857780456543, + "learning_rate": 0.00019436471787192446, + "loss": 0.3563, + "step": 5420 + }, + { + "epoch": 0.08470081737068696, + "grad_norm": 3.056812047958374, + "learning_rate": 0.00019435431875376968, + "loss": 0.3094, + "step": 5430 + }, + { + "epoch": 0.08485680414300867, + "grad_norm": 1.3599086999893188, + "learning_rate": 0.00019434391963561493, + "loss": 0.3397, + "step": 5440 + }, + { + "epoch": 0.08501279091533039, + "grad_norm": 6.885451316833496, + "learning_rate": 0.00019433352051746012, + "loss": 0.3929, + "step": 5450 + }, + { + "epoch": 0.08516877768765209, + "grad_norm": 0.5930185317993164, + "learning_rate": 0.00019432312139930537, + "loss": 0.4468, + "step": 5460 + }, + { + "epoch": 0.08532476445997379, + "grad_norm": 1.0316463708877563, + "learning_rate": 0.00019431272228115056, + "loss": 0.3909, + "step": 5470 + }, + { + "epoch": 0.0854807512322955, + "grad_norm": 1.9080007076263428, + "learning_rate": 0.0001943023231629958, + "loss": 0.4185, + "step": 5480 + }, + { + "epoch": 0.08563673800461721, + "grad_norm": 1.182185173034668, + "learning_rate": 0.000194291924044841, + "loss": 0.3496, + "step": 5490 + }, + { + "epoch": 0.08579272477693892, + "grad_norm": 1.7894487380981445, + "learning_rate": 0.00019428152492668624, + "loss": 0.2275, + "step": 5500 + }, + { + "epoch": 0.08594871154926062, + "grad_norm": 7.630679130554199, + "learning_rate": 0.00019427112580853144, + "loss": 0.3358, + "step": 5510 + }, + { + "epoch": 0.08610469832158232, + "grad_norm": 4.102541923522949, + "learning_rate": 0.00019426072669037668, + "loss": 0.3494, + "step": 5520 + }, + { + "epoch": 0.08626068509390404, + "grad_norm": 2.618604898452759, + "learning_rate": 0.00019425032757222188, + "loss": 0.2152, + "step": 5530 + }, + { + "epoch": 0.08641667186622574, + "grad_norm": 7.212046146392822, + "learning_rate": 0.00019423992845406712, + "loss": 0.4545, + "step": 5540 + }, + { + "epoch": 0.08657265863854745, + "grad_norm": 4.502930164337158, + "learning_rate": 0.00019422952933591232, + "loss": 0.484, + "step": 5550 + }, + { + "epoch": 0.08672864541086917, + "grad_norm": 1.1972264051437378, + "learning_rate": 0.00019421913021775756, + "loss": 0.2682, + "step": 5560 + }, + { + "epoch": 0.08688463218319087, + "grad_norm": 4.042606830596924, + "learning_rate": 0.00019420873109960275, + "loss": 0.299, + "step": 5570 + }, + { + "epoch": 0.08704061895551257, + "grad_norm": 1.233870506286621, + "learning_rate": 0.000194198331981448, + "loss": 0.2416, + "step": 5580 + }, + { + "epoch": 0.08719660572783428, + "grad_norm": 2.3368418216705322, + "learning_rate": 0.0001941879328632932, + "loss": 0.3265, + "step": 5590 + }, + { + "epoch": 0.08735259250015599, + "grad_norm": 1.1264872550964355, + "learning_rate": 0.00019417753374513844, + "loss": 0.3293, + "step": 5600 + }, + { + "epoch": 0.0875085792724777, + "grad_norm": 0.9745358228683472, + "learning_rate": 0.00019416713462698363, + "loss": 0.3, + "step": 5610 + }, + { + "epoch": 0.0876645660447994, + "grad_norm": 5.301463603973389, + "learning_rate": 0.00019415673550882888, + "loss": 0.5055, + "step": 5620 + }, + { + "epoch": 0.0878205528171211, + "grad_norm": 2.5696167945861816, + "learning_rate": 0.00019414633639067407, + "loss": 0.3489, + "step": 5630 + }, + { + "epoch": 0.08797653958944282, + "grad_norm": 1.5095734596252441, + "learning_rate": 0.00019413593727251932, + "loss": 0.4118, + "step": 5640 + }, + { + "epoch": 0.08813252636176452, + "grad_norm": 2.604001522064209, + "learning_rate": 0.0001941255381543645, + "loss": 0.2596, + "step": 5650 + }, + { + "epoch": 0.08828851313408623, + "grad_norm": 4.1579365730285645, + "learning_rate": 0.00019411513903620976, + "loss": 0.2804, + "step": 5660 + }, + { + "epoch": 0.08844449990640793, + "grad_norm": 1.092995524406433, + "learning_rate": 0.00019410473991805495, + "loss": 0.524, + "step": 5670 + }, + { + "epoch": 0.08860048667872965, + "grad_norm": 3.346893787384033, + "learning_rate": 0.0001940943407999002, + "loss": 0.3079, + "step": 5680 + }, + { + "epoch": 0.08875647345105135, + "grad_norm": 5.878612995147705, + "learning_rate": 0.0001940839416817454, + "loss": 0.5345, + "step": 5690 + }, + { + "epoch": 0.08891246022337305, + "grad_norm": 5.466742515563965, + "learning_rate": 0.00019407354256359064, + "loss": 0.2776, + "step": 5700 + }, + { + "epoch": 0.08906844699569476, + "grad_norm": 1.969839096069336, + "learning_rate": 0.00019406314344543583, + "loss": 0.1536, + "step": 5710 + }, + { + "epoch": 0.08922443376801648, + "grad_norm": 0.32240307331085205, + "learning_rate": 0.00019405274432728108, + "loss": 0.0561, + "step": 5720 + }, + { + "epoch": 0.08938042054033818, + "grad_norm": 2.4503121376037598, + "learning_rate": 0.00019404234520912627, + "loss": 0.457, + "step": 5730 + }, + { + "epoch": 0.08953640731265988, + "grad_norm": 2.636259078979492, + "learning_rate": 0.00019403194609097152, + "loss": 0.2433, + "step": 5740 + }, + { + "epoch": 0.0896923940849816, + "grad_norm": 1.139318585395813, + "learning_rate": 0.0001940215469728167, + "loss": 0.1501, + "step": 5750 + }, + { + "epoch": 0.0898483808573033, + "grad_norm": 3.3219454288482666, + "learning_rate": 0.00019401114785466193, + "loss": 0.4036, + "step": 5760 + }, + { + "epoch": 0.090004367629625, + "grad_norm": 5.504943370819092, + "learning_rate": 0.00019400074873650715, + "loss": 0.3269, + "step": 5770 + }, + { + "epoch": 0.09016035440194671, + "grad_norm": 0.14088541269302368, + "learning_rate": 0.00019399034961835237, + "loss": 0.4199, + "step": 5780 + }, + { + "epoch": 0.09031634117426843, + "grad_norm": 2.2649993896484375, + "learning_rate": 0.00019397995050019759, + "loss": 0.5362, + "step": 5790 + }, + { + "epoch": 0.09047232794659013, + "grad_norm": 0.7106287479400635, + "learning_rate": 0.0001939695513820428, + "loss": 0.2285, + "step": 5800 + }, + { + "epoch": 0.09062831471891183, + "grad_norm": 5.814300060272217, + "learning_rate": 0.00019395915226388803, + "loss": 0.5139, + "step": 5810 + }, + { + "epoch": 0.09078430149123354, + "grad_norm": 4.537754058837891, + "learning_rate": 0.00019394875314573325, + "loss": 0.2923, + "step": 5820 + }, + { + "epoch": 0.09094028826355526, + "grad_norm": 4.533511638641357, + "learning_rate": 0.00019393835402757846, + "loss": 0.2571, + "step": 5830 + }, + { + "epoch": 0.09109627503587696, + "grad_norm": 1.4505037069320679, + "learning_rate": 0.00019392795490942368, + "loss": 0.2398, + "step": 5840 + }, + { + "epoch": 0.09125226180819866, + "grad_norm": 2.154186248779297, + "learning_rate": 0.0001939175557912689, + "loss": 0.2248, + "step": 5850 + }, + { + "epoch": 0.09140824858052037, + "grad_norm": 1.9082036018371582, + "learning_rate": 0.00019390715667311412, + "loss": 0.4279, + "step": 5860 + }, + { + "epoch": 0.09156423535284208, + "grad_norm": 3.025155782699585, + "learning_rate": 0.00019389675755495934, + "loss": 0.322, + "step": 5870 + }, + { + "epoch": 0.09172022212516379, + "grad_norm": 3.047150135040283, + "learning_rate": 0.00019388635843680456, + "loss": 0.3582, + "step": 5880 + }, + { + "epoch": 0.09187620889748549, + "grad_norm": 2.638498067855835, + "learning_rate": 0.00019387595931864978, + "loss": 0.3079, + "step": 5890 + }, + { + "epoch": 0.0920321956698072, + "grad_norm": 0.13002626597881317, + "learning_rate": 0.000193865560200495, + "loss": 0.3355, + "step": 5900 + }, + { + "epoch": 0.09218818244212891, + "grad_norm": 5.473140716552734, + "learning_rate": 0.00019385516108234022, + "loss": 0.4944, + "step": 5910 + }, + { + "epoch": 0.09234416921445061, + "grad_norm": 1.5157524347305298, + "learning_rate": 0.00019384476196418544, + "loss": 0.2071, + "step": 5920 + }, + { + "epoch": 0.09250015598677232, + "grad_norm": 1.8442835807800293, + "learning_rate": 0.00019383436284603066, + "loss": 0.2215, + "step": 5930 + }, + { + "epoch": 0.09265614275909403, + "grad_norm": 2.5450761318206787, + "learning_rate": 0.00019382396372787588, + "loss": 0.3973, + "step": 5940 + }, + { + "epoch": 0.09281212953141574, + "grad_norm": 0.356424480676651, + "learning_rate": 0.0001938135646097211, + "loss": 0.2876, + "step": 5950 + }, + { + "epoch": 0.09296811630373744, + "grad_norm": 4.990633487701416, + "learning_rate": 0.00019380316549156632, + "loss": 0.1775, + "step": 5960 + }, + { + "epoch": 0.09312410307605914, + "grad_norm": 8.785608291625977, + "learning_rate": 0.00019379276637341154, + "loss": 0.4081, + "step": 5970 + }, + { + "epoch": 0.09328008984838086, + "grad_norm": 1.021690845489502, + "learning_rate": 0.00019378236725525676, + "loss": 0.1727, + "step": 5980 + }, + { + "epoch": 0.09343607662070257, + "grad_norm": 0.6903918385505676, + "learning_rate": 0.00019377196813710198, + "loss": 0.4199, + "step": 5990 + }, + { + "epoch": 0.09359206339302427, + "grad_norm": 6.114228248596191, + "learning_rate": 0.0001937615690189472, + "loss": 0.2502, + "step": 6000 + }, + { + "epoch": 0.09374805016534597, + "grad_norm": 4.526939868927002, + "learning_rate": 0.00019375116990079242, + "loss": 0.3677, + "step": 6010 + }, + { + "epoch": 0.09390403693766769, + "grad_norm": 3.766489028930664, + "learning_rate": 0.00019374077078263764, + "loss": 0.4365, + "step": 6020 + }, + { + "epoch": 0.0940600237099894, + "grad_norm": 0.553138256072998, + "learning_rate": 0.00019373037166448286, + "loss": 0.2203, + "step": 6030 + }, + { + "epoch": 0.0942160104823111, + "grad_norm": 8.94212818145752, + "learning_rate": 0.00019371997254632808, + "loss": 0.2823, + "step": 6040 + }, + { + "epoch": 0.09437199725463281, + "grad_norm": 4.464921951293945, + "learning_rate": 0.0001937095734281733, + "loss": 0.3549, + "step": 6050 + }, + { + "epoch": 0.09452798402695452, + "grad_norm": 0.8132085204124451, + "learning_rate": 0.00019369917431001852, + "loss": 0.4469, + "step": 6060 + }, + { + "epoch": 0.09468397079927622, + "grad_norm": 2.162436008453369, + "learning_rate": 0.00019368877519186374, + "loss": 0.2974, + "step": 6070 + }, + { + "epoch": 0.09483995757159792, + "grad_norm": 3.0433995723724365, + "learning_rate": 0.00019367837607370896, + "loss": 0.4088, + "step": 6080 + }, + { + "epoch": 0.09499594434391964, + "grad_norm": 0.16140206158161163, + "learning_rate": 0.00019366797695555418, + "loss": 0.2823, + "step": 6090 + }, + { + "epoch": 0.09515193111624135, + "grad_norm": 0.8597097992897034, + "learning_rate": 0.0001936575778373994, + "loss": 0.2961, + "step": 6100 + }, + { + "epoch": 0.09530791788856305, + "grad_norm": 0.8245694637298584, + "learning_rate": 0.00019364717871924461, + "loss": 0.1174, + "step": 6110 + }, + { + "epoch": 0.09546390466088475, + "grad_norm": 4.541038513183594, + "learning_rate": 0.00019363677960108983, + "loss": 0.2867, + "step": 6120 + }, + { + "epoch": 0.09561989143320647, + "grad_norm": 10.14047622680664, + "learning_rate": 0.00019362638048293505, + "loss": 0.7289, + "step": 6130 + }, + { + "epoch": 0.09577587820552817, + "grad_norm": 2.899733304977417, + "learning_rate": 0.00019361598136478027, + "loss": 0.3055, + "step": 6140 + }, + { + "epoch": 0.09593186497784988, + "grad_norm": 3.6699278354644775, + "learning_rate": 0.0001936055822466255, + "loss": 0.285, + "step": 6150 + }, + { + "epoch": 0.09608785175017158, + "grad_norm": 1.7877542972564697, + "learning_rate": 0.0001935951831284707, + "loss": 0.3871, + "step": 6160 + }, + { + "epoch": 0.0962438385224933, + "grad_norm": 2.966464042663574, + "learning_rate": 0.00019358478401031593, + "loss": 0.3512, + "step": 6170 + }, + { + "epoch": 0.096399825294815, + "grad_norm": 2.51312255859375, + "learning_rate": 0.00019357438489216115, + "loss": 0.3996, + "step": 6180 + }, + { + "epoch": 0.0965558120671367, + "grad_norm": 5.688634872436523, + "learning_rate": 0.00019356398577400637, + "loss": 0.3632, + "step": 6190 + }, + { + "epoch": 0.09671179883945841, + "grad_norm": 6.877967834472656, + "learning_rate": 0.0001935535866558516, + "loss": 0.3548, + "step": 6200 + }, + { + "epoch": 0.09686778561178012, + "grad_norm": 3.846853256225586, + "learning_rate": 0.0001935431875376968, + "loss": 0.3466, + "step": 6210 + }, + { + "epoch": 0.09702377238410183, + "grad_norm": 5.721917629241943, + "learning_rate": 0.00019353278841954203, + "loss": 0.3567, + "step": 6220 + }, + { + "epoch": 0.09717975915642353, + "grad_norm": 2.47615647315979, + "learning_rate": 0.00019352238930138725, + "loss": 0.358, + "step": 6230 + }, + { + "epoch": 0.09733574592874525, + "grad_norm": 0.843755304813385, + "learning_rate": 0.00019351199018323247, + "loss": 0.3825, + "step": 6240 + }, + { + "epoch": 0.09749173270106695, + "grad_norm": 2.846238136291504, + "learning_rate": 0.0001935015910650777, + "loss": 0.1483, + "step": 6250 + }, + { + "epoch": 0.09764771947338866, + "grad_norm": 4.587934494018555, + "learning_rate": 0.0001934911919469229, + "loss": 0.3881, + "step": 6260 + }, + { + "epoch": 0.09780370624571036, + "grad_norm": 2.4678304195404053, + "learning_rate": 0.00019348079282876813, + "loss": 0.3414, + "step": 6270 + }, + { + "epoch": 0.09795969301803208, + "grad_norm": 0.5573933124542236, + "learning_rate": 0.00019347039371061335, + "loss": 0.2158, + "step": 6280 + }, + { + "epoch": 0.09811567979035378, + "grad_norm": 2.4001994132995605, + "learning_rate": 0.00019345999459245857, + "loss": 0.3069, + "step": 6290 + }, + { + "epoch": 0.09827166656267548, + "grad_norm": 1.2843972444534302, + "learning_rate": 0.0001934495954743038, + "loss": 0.2112, + "step": 6300 + }, + { + "epoch": 0.09842765333499719, + "grad_norm": 6.185412406921387, + "learning_rate": 0.000193439196356149, + "loss": 0.3721, + "step": 6310 + }, + { + "epoch": 0.0985836401073189, + "grad_norm": 2.1380372047424316, + "learning_rate": 0.00019342879723799423, + "loss": 0.304, + "step": 6320 + }, + { + "epoch": 0.09873962687964061, + "grad_norm": 4.583189964294434, + "learning_rate": 0.00019341839811983945, + "loss": 0.3399, + "step": 6330 + }, + { + "epoch": 0.09889561365196231, + "grad_norm": 1.0530366897583008, + "learning_rate": 0.00019340799900168467, + "loss": 0.2154, + "step": 6340 + }, + { + "epoch": 0.09905160042428401, + "grad_norm": 7.6255202293396, + "learning_rate": 0.00019339759988352989, + "loss": 0.4958, + "step": 6350 + }, + { + "epoch": 0.09920758719660573, + "grad_norm": 9.250993728637695, + "learning_rate": 0.0001933872007653751, + "loss": 0.8084, + "step": 6360 + }, + { + "epoch": 0.09936357396892743, + "grad_norm": 1.315562129020691, + "learning_rate": 0.00019337680164722033, + "loss": 0.4946, + "step": 6370 + }, + { + "epoch": 0.09951956074124914, + "grad_norm": 0.964072048664093, + "learning_rate": 0.00019336640252906555, + "loss": 0.2992, + "step": 6380 + }, + { + "epoch": 0.09967554751357086, + "grad_norm": 2.9351625442504883, + "learning_rate": 0.00019335600341091076, + "loss": 0.3194, + "step": 6390 + }, + { + "epoch": 0.09983153428589256, + "grad_norm": 4.353033542633057, + "learning_rate": 0.00019334560429275598, + "loss": 0.4996, + "step": 6400 + }, + { + "epoch": 0.09998752105821426, + "grad_norm": 6.008812427520752, + "learning_rate": 0.0001933352051746012, + "loss": 0.5374, + "step": 6410 + }, + { + "epoch": 0.10014350783053597, + "grad_norm": 3.2766318321228027, + "learning_rate": 0.00019332480605644642, + "loss": 0.3351, + "step": 6420 + }, + { + "epoch": 0.10029949460285768, + "grad_norm": 0.8730254769325256, + "learning_rate": 0.00019331440693829164, + "loss": 0.3071, + "step": 6430 + }, + { + "epoch": 0.10045548137517939, + "grad_norm": 2.073676347732544, + "learning_rate": 0.00019330400782013686, + "loss": 0.3931, + "step": 6440 + }, + { + "epoch": 0.10061146814750109, + "grad_norm": 10.372605323791504, + "learning_rate": 0.00019329360870198208, + "loss": 0.4514, + "step": 6450 + }, + { + "epoch": 0.1007674549198228, + "grad_norm": 5.290284633636475, + "learning_rate": 0.0001932832095838273, + "loss": 0.3721, + "step": 6460 + }, + { + "epoch": 0.10092344169214451, + "grad_norm": 1.0684601068496704, + "learning_rate": 0.00019327281046567252, + "loss": 0.2702, + "step": 6470 + }, + { + "epoch": 0.10107942846446621, + "grad_norm": 1.874252200126648, + "learning_rate": 0.00019326241134751774, + "loss": 0.4048, + "step": 6480 + }, + { + "epoch": 0.10123541523678792, + "grad_norm": 1.0442904233932495, + "learning_rate": 0.00019325201222936296, + "loss": 0.0949, + "step": 6490 + }, + { + "epoch": 0.10139140200910962, + "grad_norm": 4.850639820098877, + "learning_rate": 0.00019324161311120818, + "loss": 0.5827, + "step": 6500 + }, + { + "epoch": 0.10154738878143134, + "grad_norm": 1.0578861236572266, + "learning_rate": 0.0001932312139930534, + "loss": 0.1895, + "step": 6510 + }, + { + "epoch": 0.10170337555375304, + "grad_norm": 2.8853342533111572, + "learning_rate": 0.00019322081487489862, + "loss": 0.3856, + "step": 6520 + }, + { + "epoch": 0.10185936232607475, + "grad_norm": 2.4124507904052734, + "learning_rate": 0.00019321041575674384, + "loss": 0.4219, + "step": 6530 + }, + { + "epoch": 0.10201534909839646, + "grad_norm": 3.3235878944396973, + "learning_rate": 0.00019320001663858906, + "loss": 0.4442, + "step": 6540 + }, + { + "epoch": 0.10217133587071817, + "grad_norm": 2.4012255668640137, + "learning_rate": 0.00019318961752043428, + "loss": 0.3567, + "step": 6550 + }, + { + "epoch": 0.10232732264303987, + "grad_norm": 2.174771785736084, + "learning_rate": 0.0001931792184022795, + "loss": 0.4336, + "step": 6560 + }, + { + "epoch": 0.10248330941536157, + "grad_norm": 1.151854395866394, + "learning_rate": 0.00019316881928412472, + "loss": 0.2289, + "step": 6570 + }, + { + "epoch": 0.10263929618768329, + "grad_norm": 3.706355571746826, + "learning_rate": 0.00019315842016596994, + "loss": 0.4042, + "step": 6580 + }, + { + "epoch": 0.102795282960005, + "grad_norm": 1.293565273284912, + "learning_rate": 0.00019314802104781516, + "loss": 0.3473, + "step": 6590 + }, + { + "epoch": 0.1029512697323267, + "grad_norm": 1.179175615310669, + "learning_rate": 0.00019313762192966038, + "loss": 0.2029, + "step": 6600 + }, + { + "epoch": 0.1031072565046484, + "grad_norm": 1.1723263263702393, + "learning_rate": 0.0001931272228115056, + "loss": 0.568, + "step": 6610 + }, + { + "epoch": 0.10326324327697012, + "grad_norm": 4.858260154724121, + "learning_rate": 0.00019311682369335082, + "loss": 0.3813, + "step": 6620 + }, + { + "epoch": 0.10341923004929182, + "grad_norm": 1.0854183435440063, + "learning_rate": 0.00019310642457519604, + "loss": 0.3765, + "step": 6630 + }, + { + "epoch": 0.10357521682161352, + "grad_norm": 2.6507794857025146, + "learning_rate": 0.00019309602545704126, + "loss": 0.2726, + "step": 6640 + }, + { + "epoch": 0.10373120359393523, + "grad_norm": 0.2786160111427307, + "learning_rate": 0.00019308562633888648, + "loss": 0.4565, + "step": 6650 + }, + { + "epoch": 0.10388719036625695, + "grad_norm": 3.4936556816101074, + "learning_rate": 0.00019307522722073167, + "loss": 0.25, + "step": 6660 + }, + { + "epoch": 0.10404317713857865, + "grad_norm": 2.4455699920654297, + "learning_rate": 0.00019306482810257691, + "loss": 0.4754, + "step": 6670 + }, + { + "epoch": 0.10419916391090035, + "grad_norm": 2.3813774585723877, + "learning_rate": 0.0001930544289844221, + "loss": 0.2868, + "step": 6680 + }, + { + "epoch": 0.10435515068322206, + "grad_norm": 3.8882644176483154, + "learning_rate": 0.00019304402986626735, + "loss": 0.2338, + "step": 6690 + }, + { + "epoch": 0.10451113745554377, + "grad_norm": 2.1813762187957764, + "learning_rate": 0.00019303363074811255, + "loss": 0.2803, + "step": 6700 + }, + { + "epoch": 0.10466712422786548, + "grad_norm": 4.09452486038208, + "learning_rate": 0.0001930232316299578, + "loss": 0.3288, + "step": 6710 + }, + { + "epoch": 0.10482311100018718, + "grad_norm": 2.630009412765503, + "learning_rate": 0.00019301283251180299, + "loss": 0.3096, + "step": 6720 + }, + { + "epoch": 0.1049790977725089, + "grad_norm": 1.8648737668991089, + "learning_rate": 0.00019300243339364823, + "loss": 0.5055, + "step": 6730 + }, + { + "epoch": 0.1051350845448306, + "grad_norm": 1.4091843366622925, + "learning_rate": 0.00019299203427549343, + "loss": 0.2747, + "step": 6740 + }, + { + "epoch": 0.1052910713171523, + "grad_norm": 1.8997440338134766, + "learning_rate": 0.00019298163515733867, + "loss": 0.4575, + "step": 6750 + }, + { + "epoch": 0.10544705808947401, + "grad_norm": 3.2947158813476562, + "learning_rate": 0.0001929712360391839, + "loss": 0.251, + "step": 6760 + }, + { + "epoch": 0.10560304486179573, + "grad_norm": 1.6467453241348267, + "learning_rate": 0.0001929608369210291, + "loss": 0.2899, + "step": 6770 + }, + { + "epoch": 0.10575903163411743, + "grad_norm": 1.930840015411377, + "learning_rate": 0.00019295043780287433, + "loss": 0.1265, + "step": 6780 + }, + { + "epoch": 0.10591501840643913, + "grad_norm": 9.051993370056152, + "learning_rate": 0.00019294003868471955, + "loss": 0.3849, + "step": 6790 + }, + { + "epoch": 0.10607100517876084, + "grad_norm": 2.5743913650512695, + "learning_rate": 0.00019292963956656477, + "loss": 0.2501, + "step": 6800 + }, + { + "epoch": 0.10622699195108255, + "grad_norm": 1.333158016204834, + "learning_rate": 0.00019291924044841, + "loss": 0.415, + "step": 6810 + }, + { + "epoch": 0.10638297872340426, + "grad_norm": 4.270011901855469, + "learning_rate": 0.0001929088413302552, + "loss": 0.2877, + "step": 6820 + }, + { + "epoch": 0.10653896549572596, + "grad_norm": 2.4836020469665527, + "learning_rate": 0.00019289844221210043, + "loss": 0.5365, + "step": 6830 + }, + { + "epoch": 0.10669495226804766, + "grad_norm": 0.6873673796653748, + "learning_rate": 0.00019288804309394565, + "loss": 0.3269, + "step": 6840 + }, + { + "epoch": 0.10685093904036938, + "grad_norm": 3.456326723098755, + "learning_rate": 0.00019287764397579087, + "loss": 0.3082, + "step": 6850 + }, + { + "epoch": 0.10700692581269108, + "grad_norm": 1.6058659553527832, + "learning_rate": 0.0001928672448576361, + "loss": 0.2203, + "step": 6860 + }, + { + "epoch": 0.10716291258501279, + "grad_norm": 10.284086227416992, + "learning_rate": 0.0001928568457394813, + "loss": 0.4912, + "step": 6870 + }, + { + "epoch": 0.1073188993573345, + "grad_norm": 2.4946155548095703, + "learning_rate": 0.00019284644662132653, + "loss": 0.169, + "step": 6880 + }, + { + "epoch": 0.10747488612965621, + "grad_norm": 3.717564105987549, + "learning_rate": 0.00019283604750317175, + "loss": 0.3952, + "step": 6890 + }, + { + "epoch": 0.10763087290197791, + "grad_norm": 1.3209753036499023, + "learning_rate": 0.00019282564838501697, + "loss": 0.2958, + "step": 6900 + }, + { + "epoch": 0.10778685967429961, + "grad_norm": 2.270832061767578, + "learning_rate": 0.00019281524926686219, + "loss": 0.3197, + "step": 6910 + }, + { + "epoch": 0.10794284644662133, + "grad_norm": 3.6317005157470703, + "learning_rate": 0.0001928048501487074, + "loss": 0.234, + "step": 6920 + }, + { + "epoch": 0.10809883321894304, + "grad_norm": 1.7942756414413452, + "learning_rate": 0.00019279445103055263, + "loss": 0.3435, + "step": 6930 + }, + { + "epoch": 0.10825481999126474, + "grad_norm": 0.7844222784042358, + "learning_rate": 0.00019278405191239785, + "loss": 0.3194, + "step": 6940 + }, + { + "epoch": 0.10841080676358644, + "grad_norm": 4.149482727050781, + "learning_rate": 0.00019277365279424306, + "loss": 0.2947, + "step": 6950 + }, + { + "epoch": 0.10856679353590816, + "grad_norm": 5.077116966247559, + "learning_rate": 0.00019276325367608828, + "loss": 0.3631, + "step": 6960 + }, + { + "epoch": 0.10872278030822986, + "grad_norm": 1.661909580230713, + "learning_rate": 0.0001927528545579335, + "loss": 0.2069, + "step": 6970 + }, + { + "epoch": 0.10887876708055157, + "grad_norm": 4.791439533233643, + "learning_rate": 0.00019274245543977872, + "loss": 0.3677, + "step": 6980 + }, + { + "epoch": 0.10903475385287327, + "grad_norm": 5.538137912750244, + "learning_rate": 0.00019273205632162394, + "loss": 0.2138, + "step": 6990 + }, + { + "epoch": 0.10919074062519499, + "grad_norm": 2.9961178302764893, + "learning_rate": 0.00019272165720346916, + "loss": 0.3518, + "step": 7000 + }, + { + "epoch": 0.10934672739751669, + "grad_norm": 2.286130666732788, + "learning_rate": 0.00019271125808531438, + "loss": 0.4359, + "step": 7010 + }, + { + "epoch": 0.1095027141698384, + "grad_norm": 1.5699574947357178, + "learning_rate": 0.0001927008589671596, + "loss": 0.3327, + "step": 7020 + }, + { + "epoch": 0.10965870094216011, + "grad_norm": 0.44791361689567566, + "learning_rate": 0.00019269045984900482, + "loss": 0.2382, + "step": 7030 + }, + { + "epoch": 0.10981468771448182, + "grad_norm": 1.4798070192337036, + "learning_rate": 0.00019268006073085004, + "loss": 0.3653, + "step": 7040 + }, + { + "epoch": 0.10997067448680352, + "grad_norm": 1.1877557039260864, + "learning_rate": 0.00019266966161269526, + "loss": 0.3338, + "step": 7050 + }, + { + "epoch": 0.11012666125912522, + "grad_norm": 0.3134145140647888, + "learning_rate": 0.00019265926249454048, + "loss": 0.3832, + "step": 7060 + }, + { + "epoch": 0.11028264803144694, + "grad_norm": 1.1449509859085083, + "learning_rate": 0.0001926488633763857, + "loss": 0.27, + "step": 7070 + }, + { + "epoch": 0.11043863480376864, + "grad_norm": 0.8753034472465515, + "learning_rate": 0.00019263846425823092, + "loss": 0.3869, + "step": 7080 + }, + { + "epoch": 0.11059462157609035, + "grad_norm": 4.871153831481934, + "learning_rate": 0.00019262806514007614, + "loss": 0.3679, + "step": 7090 + }, + { + "epoch": 0.11075060834841205, + "grad_norm": 2.3015692234039307, + "learning_rate": 0.00019261766602192136, + "loss": 0.2879, + "step": 7100 + }, + { + "epoch": 0.11090659512073377, + "grad_norm": 10.451770782470703, + "learning_rate": 0.00019260726690376655, + "loss": 0.4762, + "step": 7110 + }, + { + "epoch": 0.11106258189305547, + "grad_norm": 0.6290777325630188, + "learning_rate": 0.0001925968677856118, + "loss": 0.2412, + "step": 7120 + }, + { + "epoch": 0.11121856866537717, + "grad_norm": 3.2147696018218994, + "learning_rate": 0.000192586468667457, + "loss": 0.3651, + "step": 7130 + }, + { + "epoch": 0.11137455543769888, + "grad_norm": 2.770252227783203, + "learning_rate": 0.00019257606954930224, + "loss": 0.2206, + "step": 7140 + }, + { + "epoch": 0.1115305422100206, + "grad_norm": 5.446028232574463, + "learning_rate": 0.00019256567043114743, + "loss": 0.2899, + "step": 7150 + }, + { + "epoch": 0.1116865289823423, + "grad_norm": 1.5418429374694824, + "learning_rate": 0.00019255527131299268, + "loss": 0.2798, + "step": 7160 + }, + { + "epoch": 0.111842515754664, + "grad_norm": 0.7440494298934937, + "learning_rate": 0.00019254487219483787, + "loss": 0.5074, + "step": 7170 + }, + { + "epoch": 0.1119985025269857, + "grad_norm": 3.4700686931610107, + "learning_rate": 0.00019253447307668312, + "loss": 0.2211, + "step": 7180 + }, + { + "epoch": 0.11215448929930742, + "grad_norm": 3.707090139389038, + "learning_rate": 0.0001925240739585283, + "loss": 0.4275, + "step": 7190 + }, + { + "epoch": 0.11231047607162913, + "grad_norm": 4.110064506530762, + "learning_rate": 0.00019251367484037356, + "loss": 0.2144, + "step": 7200 + }, + { + "epoch": 0.11246646284395083, + "grad_norm": 2.0909435749053955, + "learning_rate": 0.00019250327572221875, + "loss": 0.2126, + "step": 7210 + }, + { + "epoch": 0.11262244961627255, + "grad_norm": 1.880205512046814, + "learning_rate": 0.000192492876604064, + "loss": 0.2941, + "step": 7220 + }, + { + "epoch": 0.11277843638859425, + "grad_norm": 2.2292237281799316, + "learning_rate": 0.0001924824774859092, + "loss": 0.3682, + "step": 7230 + }, + { + "epoch": 0.11293442316091595, + "grad_norm": 1.1748030185699463, + "learning_rate": 0.00019247207836775443, + "loss": 0.3069, + "step": 7240 + }, + { + "epoch": 0.11309040993323766, + "grad_norm": 3.5196616649627686, + "learning_rate": 0.00019246167924959963, + "loss": 0.2736, + "step": 7250 + }, + { + "epoch": 0.11324639670555937, + "grad_norm": 0.2772327959537506, + "learning_rate": 0.00019245128013144487, + "loss": 0.3013, + "step": 7260 + }, + { + "epoch": 0.11340238347788108, + "grad_norm": 3.669743537902832, + "learning_rate": 0.00019244088101329007, + "loss": 0.6027, + "step": 7270 + }, + { + "epoch": 0.11355837025020278, + "grad_norm": 0.13140790164470673, + "learning_rate": 0.0001924304818951353, + "loss": 0.1937, + "step": 7280 + }, + { + "epoch": 0.11371435702252448, + "grad_norm": 5.190703392028809, + "learning_rate": 0.0001924200827769805, + "loss": 0.228, + "step": 7290 + }, + { + "epoch": 0.1138703437948462, + "grad_norm": 0.5808380246162415, + "learning_rate": 0.00019240968365882575, + "loss": 0.5241, + "step": 7300 + }, + { + "epoch": 0.1140263305671679, + "grad_norm": 1.059551477432251, + "learning_rate": 0.00019239928454067094, + "loss": 0.1928, + "step": 7310 + }, + { + "epoch": 0.11418231733948961, + "grad_norm": 0.9926179051399231, + "learning_rate": 0.0001923888854225162, + "loss": 0.5882, + "step": 7320 + }, + { + "epoch": 0.11433830411181131, + "grad_norm": 3.8654983043670654, + "learning_rate": 0.00019237848630436138, + "loss": 0.2472, + "step": 7330 + }, + { + "epoch": 0.11449429088413303, + "grad_norm": 1.716522455215454, + "learning_rate": 0.00019236808718620663, + "loss": 0.4283, + "step": 7340 + }, + { + "epoch": 0.11465027765645473, + "grad_norm": 0.9728891849517822, + "learning_rate": 0.00019235768806805182, + "loss": 0.183, + "step": 7350 + }, + { + "epoch": 0.11480626442877644, + "grad_norm": 0.9542707204818726, + "learning_rate": 0.00019234728894989707, + "loss": 0.0839, + "step": 7360 + }, + { + "epoch": 0.11496225120109815, + "grad_norm": 0.8934445977210999, + "learning_rate": 0.00019233688983174226, + "loss": 0.3306, + "step": 7370 + }, + { + "epoch": 0.11511823797341986, + "grad_norm": 3.0257279872894287, + "learning_rate": 0.0001923264907135875, + "loss": 0.3541, + "step": 7380 + }, + { + "epoch": 0.11527422474574156, + "grad_norm": 1.0677869319915771, + "learning_rate": 0.0001923160915954327, + "loss": 0.2027, + "step": 7390 + }, + { + "epoch": 0.11543021151806326, + "grad_norm": 2.3273229598999023, + "learning_rate": 0.00019230569247727795, + "loss": 0.4166, + "step": 7400 + }, + { + "epoch": 0.11558619829038498, + "grad_norm": 0.7801164984703064, + "learning_rate": 0.00019229529335912314, + "loss": 0.2957, + "step": 7410 + }, + { + "epoch": 0.11574218506270668, + "grad_norm": 2.9796864986419678, + "learning_rate": 0.0001922848942409684, + "loss": 0.2919, + "step": 7420 + }, + { + "epoch": 0.11589817183502839, + "grad_norm": 3.8021762371063232, + "learning_rate": 0.00019227449512281358, + "loss": 0.2455, + "step": 7430 + }, + { + "epoch": 0.11605415860735009, + "grad_norm": 0.5679472088813782, + "learning_rate": 0.00019226409600465883, + "loss": 0.1924, + "step": 7440 + }, + { + "epoch": 0.11621014537967181, + "grad_norm": 2.501598596572876, + "learning_rate": 0.00019225369688650402, + "loss": 0.2746, + "step": 7450 + }, + { + "epoch": 0.11636613215199351, + "grad_norm": 9.79799747467041, + "learning_rate": 0.00019224329776834927, + "loss": 0.5196, + "step": 7460 + }, + { + "epoch": 0.11652211892431522, + "grad_norm": 1.9365442991256714, + "learning_rate": 0.00019223289865019446, + "loss": 0.4821, + "step": 7470 + }, + { + "epoch": 0.11667810569663692, + "grad_norm": 2.2277321815490723, + "learning_rate": 0.0001922224995320397, + "loss": 0.262, + "step": 7480 + }, + { + "epoch": 0.11683409246895864, + "grad_norm": 2.1314620971679688, + "learning_rate": 0.0001922121004138849, + "loss": 0.2229, + "step": 7490 + }, + { + "epoch": 0.11699007924128034, + "grad_norm": 0.2625826299190521, + "learning_rate": 0.00019220170129573014, + "loss": 0.2179, + "step": 7500 + }, + { + "epoch": 0.11714606601360204, + "grad_norm": 4.7705183029174805, + "learning_rate": 0.00019219130217757534, + "loss": 0.2437, + "step": 7510 + }, + { + "epoch": 0.11730205278592376, + "grad_norm": 2.144707679748535, + "learning_rate": 0.00019218090305942058, + "loss": 0.331, + "step": 7520 + }, + { + "epoch": 0.11745803955824546, + "grad_norm": 1.4071444272994995, + "learning_rate": 0.00019217050394126578, + "loss": 0.4584, + "step": 7530 + }, + { + "epoch": 0.11761402633056717, + "grad_norm": 1.5593106746673584, + "learning_rate": 0.00019216010482311102, + "loss": 0.5354, + "step": 7540 + }, + { + "epoch": 0.11777001310288887, + "grad_norm": 2.466290235519409, + "learning_rate": 0.00019214970570495622, + "loss": 0.3469, + "step": 7550 + }, + { + "epoch": 0.11792599987521059, + "grad_norm": 11.850781440734863, + "learning_rate": 0.00019213930658680144, + "loss": 0.2365, + "step": 7560 + }, + { + "epoch": 0.11808198664753229, + "grad_norm": 3.969515562057495, + "learning_rate": 0.00019212890746864666, + "loss": 0.2725, + "step": 7570 + }, + { + "epoch": 0.118237973419854, + "grad_norm": 2.140038251876831, + "learning_rate": 0.00019211850835049188, + "loss": 0.265, + "step": 7580 + }, + { + "epoch": 0.1183939601921757, + "grad_norm": 5.333108901977539, + "learning_rate": 0.0001921081092323371, + "loss": 0.3605, + "step": 7590 + }, + { + "epoch": 0.11854994696449742, + "grad_norm": 1.0600502490997314, + "learning_rate": 0.00019209771011418231, + "loss": 0.3654, + "step": 7600 + }, + { + "epoch": 0.11870593373681912, + "grad_norm": 0.09935326874256134, + "learning_rate": 0.00019208731099602753, + "loss": 0.1326, + "step": 7610 + }, + { + "epoch": 0.11886192050914082, + "grad_norm": 0.39839038252830505, + "learning_rate": 0.00019207691187787275, + "loss": 0.4367, + "step": 7620 + }, + { + "epoch": 0.11901790728146253, + "grad_norm": 0.08571303635835648, + "learning_rate": 0.000192066512759718, + "loss": 0.5292, + "step": 7630 + }, + { + "epoch": 0.11917389405378424, + "grad_norm": 2.0398521423339844, + "learning_rate": 0.0001920561136415632, + "loss": 0.2453, + "step": 7640 + }, + { + "epoch": 0.11932988082610595, + "grad_norm": 1.871935486793518, + "learning_rate": 0.00019204571452340844, + "loss": 0.2797, + "step": 7650 + }, + { + "epoch": 0.11948586759842765, + "grad_norm": 8.323614120483398, + "learning_rate": 0.00019203531540525363, + "loss": 0.1233, + "step": 7660 + }, + { + "epoch": 0.11964185437074935, + "grad_norm": 4.91874885559082, + "learning_rate": 0.00019202491628709888, + "loss": 0.3747, + "step": 7670 + }, + { + "epoch": 0.11979784114307107, + "grad_norm": 1.5453379154205322, + "learning_rate": 0.00019201451716894407, + "loss": 0.3069, + "step": 7680 + }, + { + "epoch": 0.11995382791539277, + "grad_norm": 0.7079108953475952, + "learning_rate": 0.00019200411805078932, + "loss": 0.2985, + "step": 7690 + }, + { + "epoch": 0.12010981468771448, + "grad_norm": 1.6853049993515015, + "learning_rate": 0.0001919937189326345, + "loss": 0.3572, + "step": 7700 + }, + { + "epoch": 0.1202658014600362, + "grad_norm": 0.4013264775276184, + "learning_rate": 0.00019198331981447976, + "loss": 0.2229, + "step": 7710 + }, + { + "epoch": 0.1204217882323579, + "grad_norm": 6.26057767868042, + "learning_rate": 0.00019197292069632495, + "loss": 0.3661, + "step": 7720 + }, + { + "epoch": 0.1205777750046796, + "grad_norm": 4.295512676239014, + "learning_rate": 0.0001919625215781702, + "loss": 0.2982, + "step": 7730 + }, + { + "epoch": 0.1207337617770013, + "grad_norm": 3.1866257190704346, + "learning_rate": 0.0001919521224600154, + "loss": 0.3566, + "step": 7740 + }, + { + "epoch": 0.12088974854932302, + "grad_norm": 3.4306435585021973, + "learning_rate": 0.00019194172334186064, + "loss": 0.403, + "step": 7750 + }, + { + "epoch": 0.12104573532164473, + "grad_norm": 5.156405448913574, + "learning_rate": 0.00019193132422370583, + "loss": 0.3088, + "step": 7760 + }, + { + "epoch": 0.12120172209396643, + "grad_norm": 2.769991397857666, + "learning_rate": 0.00019192092510555108, + "loss": 0.224, + "step": 7770 + }, + { + "epoch": 0.12135770886628813, + "grad_norm": 4.933774471282959, + "learning_rate": 0.00019191052598739627, + "loss": 0.5199, + "step": 7780 + }, + { + "epoch": 0.12151369563860985, + "grad_norm": 2.2560410499572754, + "learning_rate": 0.00019190012686924151, + "loss": 0.4144, + "step": 7790 + }, + { + "epoch": 0.12166968241093155, + "grad_norm": 1.0731757879257202, + "learning_rate": 0.0001918897277510867, + "loss": 0.2327, + "step": 7800 + }, + { + "epoch": 0.12182566918325326, + "grad_norm": 1.9298815727233887, + "learning_rate": 0.00019187932863293195, + "loss": 0.3142, + "step": 7810 + }, + { + "epoch": 0.12198165595557496, + "grad_norm": 2.4373934268951416, + "learning_rate": 0.00019186892951477715, + "loss": 0.3267, + "step": 7820 + }, + { + "epoch": 0.12213764272789668, + "grad_norm": 2.707430839538574, + "learning_rate": 0.0001918585303966224, + "loss": 0.3451, + "step": 7830 + }, + { + "epoch": 0.12229362950021838, + "grad_norm": 2.4040186405181885, + "learning_rate": 0.00019184813127846759, + "loss": 0.3833, + "step": 7840 + }, + { + "epoch": 0.12244961627254008, + "grad_norm": 1.0734840631484985, + "learning_rate": 0.00019183773216031283, + "loss": 0.3856, + "step": 7850 + }, + { + "epoch": 0.1226056030448618, + "grad_norm": 2.8232712745666504, + "learning_rate": 0.00019182733304215802, + "loss": 0.3998, + "step": 7860 + }, + { + "epoch": 0.1227615898171835, + "grad_norm": 2.6121976375579834, + "learning_rate": 0.00019181693392400327, + "loss": 0.3892, + "step": 7870 + }, + { + "epoch": 0.12291757658950521, + "grad_norm": 3.365650177001953, + "learning_rate": 0.00019180653480584846, + "loss": 0.2566, + "step": 7880 + }, + { + "epoch": 0.12307356336182691, + "grad_norm": 0.292998731136322, + "learning_rate": 0.0001917961356876937, + "loss": 0.2939, + "step": 7890 + }, + { + "epoch": 0.12322955013414863, + "grad_norm": 0.3753792941570282, + "learning_rate": 0.0001917857365695389, + "loss": 0.2036, + "step": 7900 + }, + { + "epoch": 0.12338553690647033, + "grad_norm": 0.28090208768844604, + "learning_rate": 0.00019177533745138415, + "loss": 0.2517, + "step": 7910 + }, + { + "epoch": 0.12354152367879204, + "grad_norm": 1.4824659824371338, + "learning_rate": 0.00019176493833322934, + "loss": 0.9074, + "step": 7920 + }, + { + "epoch": 0.12369751045111374, + "grad_norm": 5.32909631729126, + "learning_rate": 0.0001917545392150746, + "loss": 0.428, + "step": 7930 + }, + { + "epoch": 0.12385349722343546, + "grad_norm": 4.6401824951171875, + "learning_rate": 0.00019174414009691978, + "loss": 0.2052, + "step": 7940 + }, + { + "epoch": 0.12400948399575716, + "grad_norm": 1.4302258491516113, + "learning_rate": 0.00019173374097876503, + "loss": 0.1912, + "step": 7950 + }, + { + "epoch": 0.12416547076807886, + "grad_norm": 0.6276291608810425, + "learning_rate": 0.00019172334186061022, + "loss": 0.2532, + "step": 7960 + }, + { + "epoch": 0.12432145754040057, + "grad_norm": 0.4935452342033386, + "learning_rate": 0.00019171294274245547, + "loss": 0.2772, + "step": 7970 + }, + { + "epoch": 0.12447744431272229, + "grad_norm": 1.2430734634399414, + "learning_rate": 0.00019170254362430066, + "loss": 0.2626, + "step": 7980 + }, + { + "epoch": 0.12463343108504399, + "grad_norm": 2.41579270362854, + "learning_rate": 0.0001916921445061459, + "loss": 0.2515, + "step": 7990 + }, + { + "epoch": 0.12478941785736569, + "grad_norm": 16.254776000976562, + "learning_rate": 0.0001916817453879911, + "loss": 0.4629, + "step": 8000 + }, + { + "epoch": 0.12494540462968741, + "grad_norm": 0.9846019148826599, + "learning_rate": 0.00019167134626983632, + "loss": 0.1909, + "step": 8010 + }, + { + "epoch": 0.1251013914020091, + "grad_norm": 0.2564077079296112, + "learning_rate": 0.00019166094715168154, + "loss": 0.2582, + "step": 8020 + }, + { + "epoch": 0.12525737817433083, + "grad_norm": 4.15382719039917, + "learning_rate": 0.00019165054803352676, + "loss": 0.3407, + "step": 8030 + }, + { + "epoch": 0.12541336494665253, + "grad_norm": 1.3758265972137451, + "learning_rate": 0.00019164014891537198, + "loss": 0.368, + "step": 8040 + }, + { + "epoch": 0.12556935171897424, + "grad_norm": 0.429997980594635, + "learning_rate": 0.0001916297497972172, + "loss": 0.4058, + "step": 8050 + }, + { + "epoch": 0.12572533849129594, + "grad_norm": 1.7166649103164673, + "learning_rate": 0.00019161935067906242, + "loss": 0.4811, + "step": 8060 + }, + { + "epoch": 0.12588132526361764, + "grad_norm": 1.644769310951233, + "learning_rate": 0.00019160895156090764, + "loss": 0.2566, + "step": 8070 + }, + { + "epoch": 0.12603731203593935, + "grad_norm": 1.0035743713378906, + "learning_rate": 0.00019159855244275286, + "loss": 0.2881, + "step": 8080 + }, + { + "epoch": 0.12619329880826105, + "grad_norm": 11.098752975463867, + "learning_rate": 0.00019158815332459808, + "loss": 0.2669, + "step": 8090 + }, + { + "epoch": 0.12634928558058275, + "grad_norm": 2.0310935974121094, + "learning_rate": 0.0001915777542064433, + "loss": 0.6095, + "step": 8100 + }, + { + "epoch": 0.12650527235290449, + "grad_norm": 3.2923743724823, + "learning_rate": 0.00019156735508828852, + "loss": 0.3604, + "step": 8110 + }, + { + "epoch": 0.1266612591252262, + "grad_norm": 2.754688262939453, + "learning_rate": 0.00019155695597013374, + "loss": 0.373, + "step": 8120 + }, + { + "epoch": 0.1268172458975479, + "grad_norm": 1.6191695928573608, + "learning_rate": 0.00019154655685197896, + "loss": 0.342, + "step": 8130 + }, + { + "epoch": 0.1269732326698696, + "grad_norm": 4.7524285316467285, + "learning_rate": 0.00019153615773382417, + "loss": 0.2754, + "step": 8140 + }, + { + "epoch": 0.1271292194421913, + "grad_norm": 0.20522356033325195, + "learning_rate": 0.0001915257586156694, + "loss": 0.1422, + "step": 8150 + }, + { + "epoch": 0.127285206214513, + "grad_norm": 0.21395255625247955, + "learning_rate": 0.00019151535949751461, + "loss": 0.3286, + "step": 8160 + }, + { + "epoch": 0.1274411929868347, + "grad_norm": 1.322213888168335, + "learning_rate": 0.00019150496037935983, + "loss": 0.393, + "step": 8170 + }, + { + "epoch": 0.12759717975915644, + "grad_norm": 2.434922933578491, + "learning_rate": 0.00019149456126120505, + "loss": 0.2596, + "step": 8180 + }, + { + "epoch": 0.12775316653147814, + "grad_norm": 6.111631393432617, + "learning_rate": 0.00019148416214305027, + "loss": 0.7321, + "step": 8190 + }, + { + "epoch": 0.12790915330379984, + "grad_norm": 8.627694129943848, + "learning_rate": 0.0001914737630248955, + "loss": 0.2527, + "step": 8200 + }, + { + "epoch": 0.12806514007612155, + "grad_norm": 7.098100662231445, + "learning_rate": 0.0001914633639067407, + "loss": 0.3481, + "step": 8210 + }, + { + "epoch": 0.12822112684844325, + "grad_norm": 2.6047921180725098, + "learning_rate": 0.00019145296478858593, + "loss": 0.3985, + "step": 8220 + }, + { + "epoch": 0.12837711362076495, + "grad_norm": 2.3601999282836914, + "learning_rate": 0.00019144256567043115, + "loss": 0.3182, + "step": 8230 + }, + { + "epoch": 0.12853310039308666, + "grad_norm": 5.261406421661377, + "learning_rate": 0.00019143216655227637, + "loss": 0.2836, + "step": 8240 + }, + { + "epoch": 0.12868908716540836, + "grad_norm": 2.3208975791931152, + "learning_rate": 0.0001914217674341216, + "loss": 0.1436, + "step": 8250 + }, + { + "epoch": 0.1288450739377301, + "grad_norm": 0.5465450286865234, + "learning_rate": 0.0001914113683159668, + "loss": 0.1902, + "step": 8260 + }, + { + "epoch": 0.1290010607100518, + "grad_norm": 1.4398523569107056, + "learning_rate": 0.00019140096919781203, + "loss": 0.3946, + "step": 8270 + }, + { + "epoch": 0.1291570474823735, + "grad_norm": 0.4072510302066803, + "learning_rate": 0.00019139057007965725, + "loss": 0.2422, + "step": 8280 + }, + { + "epoch": 0.1293130342546952, + "grad_norm": 1.2237650156021118, + "learning_rate": 0.00019138017096150247, + "loss": 0.166, + "step": 8290 + }, + { + "epoch": 0.1294690210270169, + "grad_norm": 0.7359814047813416, + "learning_rate": 0.0001913697718433477, + "loss": 0.2737, + "step": 8300 + }, + { + "epoch": 0.1296250077993386, + "grad_norm": 2.178809642791748, + "learning_rate": 0.0001913593727251929, + "loss": 0.1729, + "step": 8310 + }, + { + "epoch": 0.1297809945716603, + "grad_norm": 3.504578113555908, + "learning_rate": 0.00019134897360703813, + "loss": 0.3232, + "step": 8320 + }, + { + "epoch": 0.12993698134398202, + "grad_norm": 1.777565360069275, + "learning_rate": 0.00019133857448888335, + "loss": 0.3073, + "step": 8330 + }, + { + "epoch": 0.13009296811630375, + "grad_norm": 2.342390298843384, + "learning_rate": 0.00019132817537072857, + "loss": 0.4191, + "step": 8340 + }, + { + "epoch": 0.13024895488862545, + "grad_norm": 1.5919382572174072, + "learning_rate": 0.0001913177762525738, + "loss": 0.361, + "step": 8350 + }, + { + "epoch": 0.13040494166094715, + "grad_norm": 5.767488956451416, + "learning_rate": 0.000191307377134419, + "loss": 0.2514, + "step": 8360 + }, + { + "epoch": 0.13056092843326886, + "grad_norm": 0.2066197395324707, + "learning_rate": 0.00019129697801626423, + "loss": 0.2475, + "step": 8370 + }, + { + "epoch": 0.13071691520559056, + "grad_norm": 11.444001197814941, + "learning_rate": 0.00019128657889810945, + "loss": 0.3819, + "step": 8380 + }, + { + "epoch": 0.13087290197791226, + "grad_norm": 2.607598304748535, + "learning_rate": 0.00019127617977995467, + "loss": 0.4103, + "step": 8390 + }, + { + "epoch": 0.13102888875023397, + "grad_norm": 4.767722129821777, + "learning_rate": 0.00019126578066179989, + "loss": 0.3511, + "step": 8400 + }, + { + "epoch": 0.1311848755225557, + "grad_norm": 3.6843106746673584, + "learning_rate": 0.0001912553815436451, + "loss": 0.4402, + "step": 8410 + }, + { + "epoch": 0.1313408622948774, + "grad_norm": 3.7585389614105225, + "learning_rate": 0.00019124498242549032, + "loss": 0.4182, + "step": 8420 + }, + { + "epoch": 0.1314968490671991, + "grad_norm": 3.5597290992736816, + "learning_rate": 0.00019123458330733554, + "loss": 0.3161, + "step": 8430 + }, + { + "epoch": 0.1316528358395208, + "grad_norm": 0.20346300303936005, + "learning_rate": 0.00019122418418918076, + "loss": 0.1689, + "step": 8440 + }, + { + "epoch": 0.1318088226118425, + "grad_norm": 3.307145118713379, + "learning_rate": 0.00019121378507102598, + "loss": 0.3883, + "step": 8450 + }, + { + "epoch": 0.13196480938416422, + "grad_norm": 0.09858717024326324, + "learning_rate": 0.0001912033859528712, + "loss": 0.1792, + "step": 8460 + }, + { + "epoch": 0.13212079615648592, + "grad_norm": 3.1137168407440186, + "learning_rate": 0.00019119298683471642, + "loss": 0.2061, + "step": 8470 + }, + { + "epoch": 0.13227678292880762, + "grad_norm": 2.4133315086364746, + "learning_rate": 0.00019118258771656164, + "loss": 0.2278, + "step": 8480 + }, + { + "epoch": 0.13243276970112935, + "grad_norm": 1.2268180847167969, + "learning_rate": 0.00019117218859840686, + "loss": 0.2337, + "step": 8490 + }, + { + "epoch": 0.13258875647345106, + "grad_norm": 2.3650662899017334, + "learning_rate": 0.00019116178948025208, + "loss": 0.4079, + "step": 8500 + }, + { + "epoch": 0.13274474324577276, + "grad_norm": 1.009161353111267, + "learning_rate": 0.0001911513903620973, + "loss": 0.4019, + "step": 8510 + }, + { + "epoch": 0.13290073001809447, + "grad_norm": 1.7153061628341675, + "learning_rate": 0.00019114099124394252, + "loss": 0.2673, + "step": 8520 + }, + { + "epoch": 0.13305671679041617, + "grad_norm": 3.887568950653076, + "learning_rate": 0.00019113059212578774, + "loss": 0.2689, + "step": 8530 + }, + { + "epoch": 0.13321270356273787, + "grad_norm": 1.7054911851882935, + "learning_rate": 0.00019112019300763296, + "loss": 0.4124, + "step": 8540 + }, + { + "epoch": 0.13336869033505958, + "grad_norm": 2.063857316970825, + "learning_rate": 0.00019110979388947818, + "loss": 0.2602, + "step": 8550 + }, + { + "epoch": 0.1335246771073813, + "grad_norm": 1.704604148864746, + "learning_rate": 0.0001910993947713234, + "loss": 0.4358, + "step": 8560 + }, + { + "epoch": 0.133680663879703, + "grad_norm": 5.169620513916016, + "learning_rate": 0.00019108899565316862, + "loss": 0.4096, + "step": 8570 + }, + { + "epoch": 0.1338366506520247, + "grad_norm": 1.4623101949691772, + "learning_rate": 0.00019107859653501384, + "loss": 0.1843, + "step": 8580 + }, + { + "epoch": 0.13399263742434642, + "grad_norm": 2.4411351680755615, + "learning_rate": 0.00019106819741685906, + "loss": 0.1862, + "step": 8590 + }, + { + "epoch": 0.13414862419666812, + "grad_norm": 2.140017032623291, + "learning_rate": 0.00019105779829870428, + "loss": 0.4326, + "step": 8600 + }, + { + "epoch": 0.13430461096898982, + "grad_norm": 3.320019483566284, + "learning_rate": 0.0001910473991805495, + "loss": 0.2574, + "step": 8610 + }, + { + "epoch": 0.13446059774131153, + "grad_norm": 1.0196399688720703, + "learning_rate": 0.00019103700006239472, + "loss": 0.1434, + "step": 8620 + }, + { + "epoch": 0.13461658451363323, + "grad_norm": 0.6640204191207886, + "learning_rate": 0.00019102660094423994, + "loss": 0.5926, + "step": 8630 + }, + { + "epoch": 0.13477257128595496, + "grad_norm": 2.0378332138061523, + "learning_rate": 0.00019101620182608516, + "loss": 0.202, + "step": 8640 + }, + { + "epoch": 0.13492855805827667, + "grad_norm": 0.19052188098430634, + "learning_rate": 0.00019100580270793038, + "loss": 0.3694, + "step": 8650 + }, + { + "epoch": 0.13508454483059837, + "grad_norm": 1.1991510391235352, + "learning_rate": 0.0001909954035897756, + "loss": 0.2576, + "step": 8660 + }, + { + "epoch": 0.13524053160292007, + "grad_norm": 0.26283520460128784, + "learning_rate": 0.00019098500447162082, + "loss": 0.1626, + "step": 8670 + }, + { + "epoch": 0.13539651837524178, + "grad_norm": 3.249889612197876, + "learning_rate": 0.00019097460535346604, + "loss": 0.3994, + "step": 8680 + }, + { + "epoch": 0.13555250514756348, + "grad_norm": 6.312870502471924, + "learning_rate": 0.00019096420623531126, + "loss": 0.4395, + "step": 8690 + }, + { + "epoch": 0.13570849191988518, + "grad_norm": 0.6005007028579712, + "learning_rate": 0.00019095380711715647, + "loss": 0.195, + "step": 8700 + }, + { + "epoch": 0.1358644786922069, + "grad_norm": 2.2957708835601807, + "learning_rate": 0.0001909434079990017, + "loss": 0.2567, + "step": 8710 + }, + { + "epoch": 0.13602046546452862, + "grad_norm": 2.172750949859619, + "learning_rate": 0.00019093300888084691, + "loss": 0.4404, + "step": 8720 + }, + { + "epoch": 0.13617645223685032, + "grad_norm": 6.653365135192871, + "learning_rate": 0.00019092260976269213, + "loss": 0.2268, + "step": 8730 + }, + { + "epoch": 0.13633243900917202, + "grad_norm": 2.227572441101074, + "learning_rate": 0.00019091221064453735, + "loss": 0.3324, + "step": 8740 + }, + { + "epoch": 0.13648842578149373, + "grad_norm": 0.9583300352096558, + "learning_rate": 0.00019090181152638257, + "loss": 0.3938, + "step": 8750 + }, + { + "epoch": 0.13664441255381543, + "grad_norm": 0.9899696111679077, + "learning_rate": 0.0001908914124082278, + "loss": 0.2501, + "step": 8760 + }, + { + "epoch": 0.13680039932613713, + "grad_norm": 3.927495241165161, + "learning_rate": 0.000190881013290073, + "loss": 0.2983, + "step": 8770 + }, + { + "epoch": 0.13695638609845884, + "grad_norm": 2.9593987464904785, + "learning_rate": 0.00019087061417191823, + "loss": 0.3336, + "step": 8780 + }, + { + "epoch": 0.13711237287078057, + "grad_norm": 1.9015326499938965, + "learning_rate": 0.00019086021505376345, + "loss": 0.4253, + "step": 8790 + }, + { + "epoch": 0.13726835964310227, + "grad_norm": 0.7660776376724243, + "learning_rate": 0.00019084981593560867, + "loss": 0.367, + "step": 8800 + }, + { + "epoch": 0.13742434641542398, + "grad_norm": 4.850160121917725, + "learning_rate": 0.0001908394168174539, + "loss": 0.2792, + "step": 8810 + }, + { + "epoch": 0.13758033318774568, + "grad_norm": 2.631519317626953, + "learning_rate": 0.0001908290176992991, + "loss": 0.3198, + "step": 8820 + }, + { + "epoch": 0.13773631996006738, + "grad_norm": 0.9831134676933289, + "learning_rate": 0.00019081861858114433, + "loss": 0.2242, + "step": 8830 + }, + { + "epoch": 0.13789230673238909, + "grad_norm": 2.2504160404205322, + "learning_rate": 0.00019080821946298955, + "loss": 0.2806, + "step": 8840 + }, + { + "epoch": 0.1380482935047108, + "grad_norm": 3.266126871109009, + "learning_rate": 0.00019079782034483477, + "loss": 0.2641, + "step": 8850 + }, + { + "epoch": 0.13820428027703252, + "grad_norm": 0.21948297321796417, + "learning_rate": 0.00019078742122668, + "loss": 0.6415, + "step": 8860 + }, + { + "epoch": 0.13836026704935422, + "grad_norm": 0.23968499898910522, + "learning_rate": 0.0001907770221085252, + "loss": 0.2328, + "step": 8870 + }, + { + "epoch": 0.13851625382167593, + "grad_norm": 2.194678544998169, + "learning_rate": 0.00019076662299037043, + "loss": 0.4037, + "step": 8880 + }, + { + "epoch": 0.13867224059399763, + "grad_norm": 1.9633435010910034, + "learning_rate": 0.00019075622387221565, + "loss": 0.501, + "step": 8890 + }, + { + "epoch": 0.13882822736631933, + "grad_norm": 1.5285439491271973, + "learning_rate": 0.00019074582475406087, + "loss": 0.1977, + "step": 8900 + }, + { + "epoch": 0.13898421413864104, + "grad_norm": 0.5968760848045349, + "learning_rate": 0.0001907354256359061, + "loss": 0.2279, + "step": 8910 + }, + { + "epoch": 0.13914020091096274, + "grad_norm": 2.4202139377593994, + "learning_rate": 0.0001907250265177513, + "loss": 0.4585, + "step": 8920 + }, + { + "epoch": 0.13929618768328444, + "grad_norm": 6.815012454986572, + "learning_rate": 0.0001907146273995965, + "loss": 0.2049, + "step": 8930 + }, + { + "epoch": 0.13945217445560618, + "grad_norm": 2.510038375854492, + "learning_rate": 0.00019070422828144175, + "loss": 0.4618, + "step": 8940 + }, + { + "epoch": 0.13960816122792788, + "grad_norm": 5.081661224365234, + "learning_rate": 0.00019069382916328697, + "loss": 0.5324, + "step": 8950 + }, + { + "epoch": 0.13976414800024958, + "grad_norm": 0.5944671630859375, + "learning_rate": 0.00019068343004513219, + "loss": 0.3111, + "step": 8960 + }, + { + "epoch": 0.1399201347725713, + "grad_norm": 4.6881022453308105, + "learning_rate": 0.0001906730309269774, + "loss": 0.2048, + "step": 8970 + }, + { + "epoch": 0.140076121544893, + "grad_norm": 2.042879104614258, + "learning_rate": 0.00019066263180882262, + "loss": 0.271, + "step": 8980 + }, + { + "epoch": 0.1402321083172147, + "grad_norm": 1.1326006650924683, + "learning_rate": 0.00019065223269066784, + "loss": 0.1859, + "step": 8990 + }, + { + "epoch": 0.1403880950895364, + "grad_norm": 2.0956993103027344, + "learning_rate": 0.00019064183357251306, + "loss": 0.1465, + "step": 9000 + }, + { + "epoch": 0.14054408186185813, + "grad_norm": 1.8095557689666748, + "learning_rate": 0.00019063143445435828, + "loss": 0.2947, + "step": 9010 + }, + { + "epoch": 0.14070006863417983, + "grad_norm": 2.5442774295806885, + "learning_rate": 0.0001906210353362035, + "loss": 0.3196, + "step": 9020 + }, + { + "epoch": 0.14085605540650153, + "grad_norm": 3.7610888481140137, + "learning_rate": 0.00019061063621804872, + "loss": 0.2794, + "step": 9030 + }, + { + "epoch": 0.14101204217882324, + "grad_norm": 0.2639630138874054, + "learning_rate": 0.00019060023709989394, + "loss": 0.339, + "step": 9040 + }, + { + "epoch": 0.14116802895114494, + "grad_norm": 1.3571760654449463, + "learning_rate": 0.00019058983798173916, + "loss": 0.295, + "step": 9050 + }, + { + "epoch": 0.14132401572346664, + "grad_norm": 1.9788237810134888, + "learning_rate": 0.00019057943886358438, + "loss": 0.145, + "step": 9060 + }, + { + "epoch": 0.14148000249578835, + "grad_norm": 1.98755943775177, + "learning_rate": 0.0001905690397454296, + "loss": 0.5241, + "step": 9070 + }, + { + "epoch": 0.14163598926811005, + "grad_norm": 11.126473426818848, + "learning_rate": 0.00019055864062727482, + "loss": 0.2084, + "step": 9080 + }, + { + "epoch": 0.14179197604043178, + "grad_norm": 6.048513412475586, + "learning_rate": 0.00019054824150912004, + "loss": 0.4383, + "step": 9090 + }, + { + "epoch": 0.1419479628127535, + "grad_norm": 6.858469486236572, + "learning_rate": 0.00019053784239096526, + "loss": 0.5582, + "step": 9100 + }, + { + "epoch": 0.1421039495850752, + "grad_norm": 0.4861454665660858, + "learning_rate": 0.00019052744327281048, + "loss": 0.2774, + "step": 9110 + }, + { + "epoch": 0.1422599363573969, + "grad_norm": 3.1820757389068604, + "learning_rate": 0.0001905170441546557, + "loss": 0.2467, + "step": 9120 + }, + { + "epoch": 0.1424159231297186, + "grad_norm": 1.196014642715454, + "learning_rate": 0.00019050664503650092, + "loss": 0.2698, + "step": 9130 + }, + { + "epoch": 0.1425719099020403, + "grad_norm": 0.511761486530304, + "learning_rate": 0.00019049624591834614, + "loss": 0.1662, + "step": 9140 + }, + { + "epoch": 0.142727896674362, + "grad_norm": 4.00806188583374, + "learning_rate": 0.00019048584680019136, + "loss": 0.4401, + "step": 9150 + }, + { + "epoch": 0.14288388344668373, + "grad_norm": 1.0736618041992188, + "learning_rate": 0.00019047544768203658, + "loss": 0.239, + "step": 9160 + }, + { + "epoch": 0.14303987021900544, + "grad_norm": 3.9986984729766846, + "learning_rate": 0.0001904650485638818, + "loss": 0.4237, + "step": 9170 + }, + { + "epoch": 0.14319585699132714, + "grad_norm": 0.895958423614502, + "learning_rate": 0.00019045464944572702, + "loss": 0.2857, + "step": 9180 + }, + { + "epoch": 0.14335184376364885, + "grad_norm": 2.562671661376953, + "learning_rate": 0.00019044425032757224, + "loss": 0.3014, + "step": 9190 + }, + { + "epoch": 0.14350783053597055, + "grad_norm": 5.249929904937744, + "learning_rate": 0.00019043385120941746, + "loss": 0.3175, + "step": 9200 + }, + { + "epoch": 0.14366381730829225, + "grad_norm": 0.5497576594352722, + "learning_rate": 0.00019042345209126268, + "loss": 0.2828, + "step": 9210 + }, + { + "epoch": 0.14381980408061396, + "grad_norm": 3.966878890991211, + "learning_rate": 0.0001904130529731079, + "loss": 0.4266, + "step": 9220 + }, + { + "epoch": 0.14397579085293566, + "grad_norm": 0.5906422734260559, + "learning_rate": 0.00019040265385495312, + "loss": 0.2646, + "step": 9230 + }, + { + "epoch": 0.1441317776252574, + "grad_norm": 2.217254877090454, + "learning_rate": 0.00019039225473679834, + "loss": 0.1863, + "step": 9240 + }, + { + "epoch": 0.1442877643975791, + "grad_norm": 1.0814603567123413, + "learning_rate": 0.00019038185561864356, + "loss": 0.3919, + "step": 9250 + }, + { + "epoch": 0.1444437511699008, + "grad_norm": 3.348808765411377, + "learning_rate": 0.00019037145650048877, + "loss": 0.3517, + "step": 9260 + }, + { + "epoch": 0.1445997379422225, + "grad_norm": 2.634974479675293, + "learning_rate": 0.000190361057382334, + "loss": 0.332, + "step": 9270 + }, + { + "epoch": 0.1447557247145442, + "grad_norm": 5.290812969207764, + "learning_rate": 0.00019035065826417921, + "loss": 0.4049, + "step": 9280 + }, + { + "epoch": 0.1449117114868659, + "grad_norm": 1.3467392921447754, + "learning_rate": 0.00019034025914602443, + "loss": 0.4742, + "step": 9290 + }, + { + "epoch": 0.1450676982591876, + "grad_norm": 1.3368233442306519, + "learning_rate": 0.00019032986002786965, + "loss": 0.2514, + "step": 9300 + }, + { + "epoch": 0.14522368503150931, + "grad_norm": 0.37425053119659424, + "learning_rate": 0.00019031946090971487, + "loss": 0.2366, + "step": 9310 + }, + { + "epoch": 0.14537967180383105, + "grad_norm": 0.47371023893356323, + "learning_rate": 0.0001903090617915601, + "loss": 0.2965, + "step": 9320 + }, + { + "epoch": 0.14553565857615275, + "grad_norm": 1.9556888341903687, + "learning_rate": 0.0001902986626734053, + "loss": 0.2265, + "step": 9330 + }, + { + "epoch": 0.14569164534847445, + "grad_norm": 1.1136163473129272, + "learning_rate": 0.00019028826355525053, + "loss": 0.2852, + "step": 9340 + }, + { + "epoch": 0.14584763212079616, + "grad_norm": 1.7750933170318604, + "learning_rate": 0.00019027786443709575, + "loss": 0.2825, + "step": 9350 + }, + { + "epoch": 0.14600361889311786, + "grad_norm": 1.3628592491149902, + "learning_rate": 0.00019026746531894097, + "loss": 0.5666, + "step": 9360 + }, + { + "epoch": 0.14615960566543956, + "grad_norm": 2.016136646270752, + "learning_rate": 0.0001902570662007862, + "loss": 0.4292, + "step": 9370 + }, + { + "epoch": 0.14631559243776127, + "grad_norm": 1.0013788938522339, + "learning_rate": 0.00019024666708263138, + "loss": 0.3201, + "step": 9380 + }, + { + "epoch": 0.146471579210083, + "grad_norm": 2.918337106704712, + "learning_rate": 0.00019023626796447663, + "loss": 0.313, + "step": 9390 + }, + { + "epoch": 0.1466275659824047, + "grad_norm": 2.9638941287994385, + "learning_rate": 0.00019022586884632182, + "loss": 0.4592, + "step": 9400 + }, + { + "epoch": 0.1467835527547264, + "grad_norm": 2.4582252502441406, + "learning_rate": 0.00019021546972816707, + "loss": 0.2585, + "step": 9410 + }, + { + "epoch": 0.1469395395270481, + "grad_norm": 4.470787525177002, + "learning_rate": 0.00019020507061001226, + "loss": 0.4089, + "step": 9420 + }, + { + "epoch": 0.1470955262993698, + "grad_norm": 1.9958068132400513, + "learning_rate": 0.0001901946714918575, + "loss": 0.6056, + "step": 9430 + }, + { + "epoch": 0.14725151307169151, + "grad_norm": 0.7107248306274414, + "learning_rate": 0.0001901842723737027, + "loss": 0.3247, + "step": 9440 + }, + { + "epoch": 0.14740749984401322, + "grad_norm": 3.0590286254882812, + "learning_rate": 0.00019017387325554795, + "loss": 0.1213, + "step": 9450 + }, + { + "epoch": 0.14756348661633492, + "grad_norm": 1.6952658891677856, + "learning_rate": 0.00019016347413739314, + "loss": 0.1593, + "step": 9460 + }, + { + "epoch": 0.14771947338865665, + "grad_norm": 5.949721336364746, + "learning_rate": 0.0001901530750192384, + "loss": 0.271, + "step": 9470 + }, + { + "epoch": 0.14787546016097836, + "grad_norm": 0.35458824038505554, + "learning_rate": 0.00019014267590108358, + "loss": 0.1206, + "step": 9480 + }, + { + "epoch": 0.14803144693330006, + "grad_norm": 1.2978142499923706, + "learning_rate": 0.00019013227678292883, + "loss": 0.4097, + "step": 9490 + }, + { + "epoch": 0.14818743370562176, + "grad_norm": 2.7621631622314453, + "learning_rate": 0.00019012187766477402, + "loss": 0.3056, + "step": 9500 + }, + { + "epoch": 0.14834342047794347, + "grad_norm": 1.4229602813720703, + "learning_rate": 0.00019011147854661927, + "loss": 0.2489, + "step": 9510 + }, + { + "epoch": 0.14849940725026517, + "grad_norm": 1.638088345527649, + "learning_rate": 0.00019010107942846446, + "loss": 0.3914, + "step": 9520 + }, + { + "epoch": 0.14865539402258687, + "grad_norm": 1.4291104078292847, + "learning_rate": 0.0001900906803103097, + "loss": 0.2537, + "step": 9530 + }, + { + "epoch": 0.1488113807949086, + "grad_norm": 1.6094496250152588, + "learning_rate": 0.0001900802811921549, + "loss": 0.258, + "step": 9540 + }, + { + "epoch": 0.1489673675672303, + "grad_norm": 5.987236976623535, + "learning_rate": 0.00019006988207400014, + "loss": 0.2963, + "step": 9550 + }, + { + "epoch": 0.149123354339552, + "grad_norm": 4.741234302520752, + "learning_rate": 0.00019005948295584534, + "loss": 0.2403, + "step": 9560 + }, + { + "epoch": 0.14927934111187371, + "grad_norm": 1.7254575490951538, + "learning_rate": 0.00019004908383769058, + "loss": 0.2019, + "step": 9570 + }, + { + "epoch": 0.14943532788419542, + "grad_norm": 0.6806448698043823, + "learning_rate": 0.00019003868471953578, + "loss": 0.3646, + "step": 9580 + }, + { + "epoch": 0.14959131465651712, + "grad_norm": 1.5645960569381714, + "learning_rate": 0.00019002828560138102, + "loss": 0.363, + "step": 9590 + }, + { + "epoch": 0.14974730142883882, + "grad_norm": 3.121180534362793, + "learning_rate": 0.00019001788648322622, + "loss": 0.4002, + "step": 9600 + }, + { + "epoch": 0.14990328820116053, + "grad_norm": 4.279516696929932, + "learning_rate": 0.00019000748736507146, + "loss": 0.1717, + "step": 9610 + }, + { + "epoch": 0.15005927497348226, + "grad_norm": 0.7724094986915588, + "learning_rate": 0.00018999708824691665, + "loss": 0.2195, + "step": 9620 + }, + { + "epoch": 0.15021526174580396, + "grad_norm": 0.4113433361053467, + "learning_rate": 0.0001899866891287619, + "loss": 0.2139, + "step": 9630 + }, + { + "epoch": 0.15037124851812567, + "grad_norm": 0.6473997235298157, + "learning_rate": 0.0001899762900106071, + "loss": 0.4487, + "step": 9640 + }, + { + "epoch": 0.15052723529044737, + "grad_norm": 0.9015470147132874, + "learning_rate": 0.00018996589089245234, + "loss": 0.2524, + "step": 9650 + }, + { + "epoch": 0.15068322206276907, + "grad_norm": 3.411315679550171, + "learning_rate": 0.00018995549177429753, + "loss": 0.438, + "step": 9660 + }, + { + "epoch": 0.15083920883509078, + "grad_norm": 3.1558773517608643, + "learning_rate": 0.00018994509265614278, + "loss": 0.3188, + "step": 9670 + }, + { + "epoch": 0.15099519560741248, + "grad_norm": 1.3154019117355347, + "learning_rate": 0.00018993469353798797, + "loss": 0.3454, + "step": 9680 + }, + { + "epoch": 0.1511511823797342, + "grad_norm": 0.18959377706050873, + "learning_rate": 0.00018992429441983322, + "loss": 0.2291, + "step": 9690 + }, + { + "epoch": 0.15130716915205591, + "grad_norm": 3.3683054447174072, + "learning_rate": 0.0001899138953016784, + "loss": 0.5085, + "step": 9700 + }, + { + "epoch": 0.15146315592437762, + "grad_norm": 1.4889847040176392, + "learning_rate": 0.00018990349618352366, + "loss": 0.1919, + "step": 9710 + }, + { + "epoch": 0.15161914269669932, + "grad_norm": 3.172668933868408, + "learning_rate": 0.00018989309706536885, + "loss": 0.397, + "step": 9720 + }, + { + "epoch": 0.15177512946902103, + "grad_norm": 3.7615721225738525, + "learning_rate": 0.0001898826979472141, + "loss": 0.4941, + "step": 9730 + }, + { + "epoch": 0.15193111624134273, + "grad_norm": 2.056947946548462, + "learning_rate": 0.0001898722988290593, + "loss": 0.3971, + "step": 9740 + }, + { + "epoch": 0.15208710301366443, + "grad_norm": 1.5753828287124634, + "learning_rate": 0.00018986189971090454, + "loss": 0.4617, + "step": 9750 + }, + { + "epoch": 0.15224308978598614, + "grad_norm": 8.168540000915527, + "learning_rate": 0.00018985150059274973, + "loss": 0.2655, + "step": 9760 + }, + { + "epoch": 0.15239907655830787, + "grad_norm": 3.8417530059814453, + "learning_rate": 0.00018984110147459498, + "loss": 0.3281, + "step": 9770 + }, + { + "epoch": 0.15255506333062957, + "grad_norm": 2.040729522705078, + "learning_rate": 0.00018983070235644017, + "loss": 0.3449, + "step": 9780 + }, + { + "epoch": 0.15271105010295127, + "grad_norm": 2.696732997894287, + "learning_rate": 0.00018982030323828542, + "loss": 0.2323, + "step": 9790 + }, + { + "epoch": 0.15286703687527298, + "grad_norm": 0.111504465341568, + "learning_rate": 0.0001898099041201306, + "loss": 0.2174, + "step": 9800 + }, + { + "epoch": 0.15302302364759468, + "grad_norm": 2.0092639923095703, + "learning_rate": 0.00018979950500197585, + "loss": 0.2481, + "step": 9810 + }, + { + "epoch": 0.15317901041991638, + "grad_norm": 2.0963873863220215, + "learning_rate": 0.00018978910588382107, + "loss": 0.2394, + "step": 9820 + }, + { + "epoch": 0.1533349971922381, + "grad_norm": 1.9664541482925415, + "learning_rate": 0.00018977870676566627, + "loss": 0.3264, + "step": 9830 + }, + { + "epoch": 0.15349098396455982, + "grad_norm": 2.2150211334228516, + "learning_rate": 0.00018976830764751151, + "loss": 0.4045, + "step": 9840 + }, + { + "epoch": 0.15364697073688152, + "grad_norm": 0.4570525586605072, + "learning_rate": 0.0001897579085293567, + "loss": 0.1828, + "step": 9850 + }, + { + "epoch": 0.15380295750920323, + "grad_norm": 8.73476505279541, + "learning_rate": 0.00018974750941120195, + "loss": 0.3996, + "step": 9860 + }, + { + "epoch": 0.15395894428152493, + "grad_norm": 0.4760019779205322, + "learning_rate": 0.00018973711029304715, + "loss": 0.2336, + "step": 9870 + }, + { + "epoch": 0.15411493105384663, + "grad_norm": 4.7811808586120605, + "learning_rate": 0.0001897267111748924, + "loss": 0.4498, + "step": 9880 + }, + { + "epoch": 0.15427091782616834, + "grad_norm": 7.987946510314941, + "learning_rate": 0.00018971631205673758, + "loss": 0.2823, + "step": 9890 + }, + { + "epoch": 0.15442690459849004, + "grad_norm": 1.7480034828186035, + "learning_rate": 0.00018970591293858283, + "loss": 0.3611, + "step": 9900 + }, + { + "epoch": 0.15458289137081174, + "grad_norm": 3.877750873565674, + "learning_rate": 0.00018969551382042802, + "loss": 0.2978, + "step": 9910 + }, + { + "epoch": 0.15473887814313347, + "grad_norm": 2.572160243988037, + "learning_rate": 0.00018968511470227327, + "loss": 0.3594, + "step": 9920 + }, + { + "epoch": 0.15489486491545518, + "grad_norm": 2.632244348526001, + "learning_rate": 0.00018967471558411846, + "loss": 0.3699, + "step": 9930 + }, + { + "epoch": 0.15505085168777688, + "grad_norm": 3.1230275630950928, + "learning_rate": 0.0001896643164659637, + "loss": 0.2424, + "step": 9940 + }, + { + "epoch": 0.15520683846009858, + "grad_norm": 2.4213674068450928, + "learning_rate": 0.0001896539173478089, + "loss": 0.252, + "step": 9950 + }, + { + "epoch": 0.1553628252324203, + "grad_norm": 1.5046018362045288, + "learning_rate": 0.00018964351822965415, + "loss": 0.4245, + "step": 9960 + }, + { + "epoch": 0.155518812004742, + "grad_norm": 2.1277291774749756, + "learning_rate": 0.00018963311911149934, + "loss": 0.3583, + "step": 9970 + }, + { + "epoch": 0.1556747987770637, + "grad_norm": 0.7622117400169373, + "learning_rate": 0.0001896227199933446, + "loss": 0.2141, + "step": 9980 + }, + { + "epoch": 0.15583078554938543, + "grad_norm": 7.480571746826172, + "learning_rate": 0.00018961232087518978, + "loss": 0.4933, + "step": 9990 + }, + { + "epoch": 0.15598677232170713, + "grad_norm": 3.018193006515503, + "learning_rate": 0.00018960192175703503, + "loss": 0.258, + "step": 10000 + }, + { + "epoch": 0.15614275909402883, + "grad_norm": 1.8125685453414917, + "learning_rate": 0.00018959152263888022, + "loss": 0.4243, + "step": 10010 + }, + { + "epoch": 0.15629874586635054, + "grad_norm": 4.893161773681641, + "learning_rate": 0.00018958112352072547, + "loss": 0.3121, + "step": 10020 + }, + { + "epoch": 0.15645473263867224, + "grad_norm": 1.7751197814941406, + "learning_rate": 0.00018957072440257066, + "loss": 0.1687, + "step": 10030 + }, + { + "epoch": 0.15661071941099394, + "grad_norm": 0.6714029908180237, + "learning_rate": 0.0001895603252844159, + "loss": 0.2726, + "step": 10040 + }, + { + "epoch": 0.15676670618331565, + "grad_norm": 1.9242407083511353, + "learning_rate": 0.0001895499261662611, + "loss": 0.5891, + "step": 10050 + }, + { + "epoch": 0.15692269295563735, + "grad_norm": 1.1395220756530762, + "learning_rate": 0.00018953952704810635, + "loss": 0.308, + "step": 10060 + }, + { + "epoch": 0.15707867972795908, + "grad_norm": 1.6830555200576782, + "learning_rate": 0.00018952912792995154, + "loss": 0.1302, + "step": 10070 + }, + { + "epoch": 0.15723466650028078, + "grad_norm": 0.1671249121427536, + "learning_rate": 0.00018951872881179679, + "loss": 0.3607, + "step": 10080 + }, + { + "epoch": 0.1573906532726025, + "grad_norm": 7.244504928588867, + "learning_rate": 0.00018950832969364198, + "loss": 0.2414, + "step": 10090 + }, + { + "epoch": 0.1575466400449242, + "grad_norm": 4.040239334106445, + "learning_rate": 0.00018949793057548722, + "loss": 0.3737, + "step": 10100 + }, + { + "epoch": 0.1577026268172459, + "grad_norm": 3.4039885997772217, + "learning_rate": 0.00018948753145733242, + "loss": 0.2634, + "step": 10110 + }, + { + "epoch": 0.1578586135895676, + "grad_norm": 1.2193446159362793, + "learning_rate": 0.00018947713233917766, + "loss": 0.2776, + "step": 10120 + }, + { + "epoch": 0.1580146003618893, + "grad_norm": 0.21849192678928375, + "learning_rate": 0.00018946673322102286, + "loss": 0.0927, + "step": 10130 + }, + { + "epoch": 0.15817058713421103, + "grad_norm": 1.1582919359207153, + "learning_rate": 0.0001894563341028681, + "loss": 0.5419, + "step": 10140 + }, + { + "epoch": 0.15832657390653274, + "grad_norm": 2.9518587589263916, + "learning_rate": 0.0001894459349847133, + "loss": 0.2916, + "step": 10150 + }, + { + "epoch": 0.15848256067885444, + "grad_norm": 1.1647896766662598, + "learning_rate": 0.00018943553586655854, + "loss": 0.1798, + "step": 10160 + }, + { + "epoch": 0.15863854745117614, + "grad_norm": 5.562665939331055, + "learning_rate": 0.00018942513674840373, + "loss": 0.2309, + "step": 10170 + }, + { + "epoch": 0.15879453422349785, + "grad_norm": 0.30232521891593933, + "learning_rate": 0.00018941473763024898, + "loss": 0.2568, + "step": 10180 + }, + { + "epoch": 0.15895052099581955, + "grad_norm": 2.403820753097534, + "learning_rate": 0.00018940433851209417, + "loss": 0.3609, + "step": 10190 + }, + { + "epoch": 0.15910650776814125, + "grad_norm": 2.416297197341919, + "learning_rate": 0.00018939393939393942, + "loss": 0.5599, + "step": 10200 + }, + { + "epoch": 0.15926249454046296, + "grad_norm": 0.4364621043205261, + "learning_rate": 0.0001893835402757846, + "loss": 0.3156, + "step": 10210 + }, + { + "epoch": 0.1594184813127847, + "grad_norm": 0.631258487701416, + "learning_rate": 0.00018937314115762986, + "loss": 0.2733, + "step": 10220 + }, + { + "epoch": 0.1595744680851064, + "grad_norm": 4.192949295043945, + "learning_rate": 0.00018936274203947505, + "loss": 0.1811, + "step": 10230 + }, + { + "epoch": 0.1597304548574281, + "grad_norm": 3.126837968826294, + "learning_rate": 0.0001893523429213203, + "loss": 0.4182, + "step": 10240 + }, + { + "epoch": 0.1598864416297498, + "grad_norm": 2.752687931060791, + "learning_rate": 0.0001893419438031655, + "loss": 0.2851, + "step": 10250 + }, + { + "epoch": 0.1600424284020715, + "grad_norm": 0.3277892470359802, + "learning_rate": 0.00018933154468501074, + "loss": 0.3376, + "step": 10260 + }, + { + "epoch": 0.1601984151743932, + "grad_norm": 1.5464520454406738, + "learning_rate": 0.00018932114556685593, + "loss": 0.3036, + "step": 10270 + }, + { + "epoch": 0.1603544019467149, + "grad_norm": 1.2205153703689575, + "learning_rate": 0.00018931074644870115, + "loss": 0.372, + "step": 10280 + }, + { + "epoch": 0.1605103887190366, + "grad_norm": 1.481124997138977, + "learning_rate": 0.00018930034733054637, + "loss": 0.3517, + "step": 10290 + }, + { + "epoch": 0.16066637549135834, + "grad_norm": 3.1430208683013916, + "learning_rate": 0.0001892899482123916, + "loss": 0.4855, + "step": 10300 + }, + { + "epoch": 0.16082236226368005, + "grad_norm": 3.9420111179351807, + "learning_rate": 0.0001892795490942368, + "loss": 0.3389, + "step": 10310 + }, + { + "epoch": 0.16097834903600175, + "grad_norm": 0.3103383183479309, + "learning_rate": 0.00018926914997608203, + "loss": 0.3516, + "step": 10320 + }, + { + "epoch": 0.16113433580832345, + "grad_norm": 3.0402798652648926, + "learning_rate": 0.00018925875085792725, + "loss": 0.4332, + "step": 10330 + }, + { + "epoch": 0.16129032258064516, + "grad_norm": 2.537154197692871, + "learning_rate": 0.00018924835173977247, + "loss": 0.2431, + "step": 10340 + }, + { + "epoch": 0.16144630935296686, + "grad_norm": 5.181357383728027, + "learning_rate": 0.0001892379526216177, + "loss": 0.1811, + "step": 10350 + }, + { + "epoch": 0.16160229612528856, + "grad_norm": 4.332000732421875, + "learning_rate": 0.0001892275535034629, + "loss": 0.5327, + "step": 10360 + }, + { + "epoch": 0.1617582828976103, + "grad_norm": 1.5739328861236572, + "learning_rate": 0.00018921715438530813, + "loss": 0.2806, + "step": 10370 + }, + { + "epoch": 0.161914269669932, + "grad_norm": 1.327676773071289, + "learning_rate": 0.00018920675526715335, + "loss": 0.2399, + "step": 10380 + }, + { + "epoch": 0.1620702564422537, + "grad_norm": 1.5268476009368896, + "learning_rate": 0.00018919635614899857, + "loss": 0.133, + "step": 10390 + }, + { + "epoch": 0.1622262432145754, + "grad_norm": 4.602615833282471, + "learning_rate": 0.0001891859570308438, + "loss": 0.5869, + "step": 10400 + }, + { + "epoch": 0.1623822299868971, + "grad_norm": 2.0539729595184326, + "learning_rate": 0.000189175557912689, + "loss": 0.3596, + "step": 10410 + }, + { + "epoch": 0.1625382167592188, + "grad_norm": 2.4833102226257324, + "learning_rate": 0.00018916515879453423, + "loss": 0.2883, + "step": 10420 + }, + { + "epoch": 0.16269420353154052, + "grad_norm": 1.984438180923462, + "learning_rate": 0.00018915475967637945, + "loss": 0.2581, + "step": 10430 + }, + { + "epoch": 0.16285019030386222, + "grad_norm": 0.4698280692100525, + "learning_rate": 0.00018914436055822467, + "loss": 0.169, + "step": 10440 + }, + { + "epoch": 0.16300617707618395, + "grad_norm": 0.044326093047857285, + "learning_rate": 0.00018913396144006988, + "loss": 0.5516, + "step": 10450 + }, + { + "epoch": 0.16316216384850565, + "grad_norm": 3.2864580154418945, + "learning_rate": 0.0001891235623219151, + "loss": 0.3672, + "step": 10460 + }, + { + "epoch": 0.16331815062082736, + "grad_norm": 3.6366145610809326, + "learning_rate": 0.00018911316320376032, + "loss": 0.2208, + "step": 10470 + }, + { + "epoch": 0.16347413739314906, + "grad_norm": 2.1250436305999756, + "learning_rate": 0.00018910276408560554, + "loss": 0.3338, + "step": 10480 + }, + { + "epoch": 0.16363012416547076, + "grad_norm": 1.1730015277862549, + "learning_rate": 0.00018909236496745076, + "loss": 0.311, + "step": 10490 + }, + { + "epoch": 0.16378611093779247, + "grad_norm": 1.017413854598999, + "learning_rate": 0.00018908196584929598, + "loss": 0.1516, + "step": 10500 + }, + { + "epoch": 0.16394209771011417, + "grad_norm": 0.03040853887796402, + "learning_rate": 0.0001890715667311412, + "loss": 0.2375, + "step": 10510 + }, + { + "epoch": 0.1640980844824359, + "grad_norm": 3.654688596725464, + "learning_rate": 0.00018906116761298642, + "loss": 0.6085, + "step": 10520 + }, + { + "epoch": 0.1642540712547576, + "grad_norm": 1.279167652130127, + "learning_rate": 0.00018905076849483164, + "loss": 0.1246, + "step": 10530 + }, + { + "epoch": 0.1644100580270793, + "grad_norm": 1.1471649408340454, + "learning_rate": 0.00018904036937667686, + "loss": 0.4182, + "step": 10540 + }, + { + "epoch": 0.164566044799401, + "grad_norm": 2.241262674331665, + "learning_rate": 0.00018902997025852208, + "loss": 0.4201, + "step": 10550 + }, + { + "epoch": 0.16472203157172272, + "grad_norm": 0.7743698954582214, + "learning_rate": 0.0001890195711403673, + "loss": 0.1443, + "step": 10560 + }, + { + "epoch": 0.16487801834404442, + "grad_norm": 2.622305393218994, + "learning_rate": 0.00018900917202221252, + "loss": 0.4357, + "step": 10570 + }, + { + "epoch": 0.16503400511636612, + "grad_norm": 2.075713634490967, + "learning_rate": 0.00018899877290405774, + "loss": 0.4837, + "step": 10580 + }, + { + "epoch": 0.16518999188868783, + "grad_norm": 3.480304002761841, + "learning_rate": 0.00018898837378590296, + "loss": 0.1797, + "step": 10590 + }, + { + "epoch": 0.16534597866100956, + "grad_norm": 0.35536032915115356, + "learning_rate": 0.00018897797466774818, + "loss": 0.134, + "step": 10600 + }, + { + "epoch": 0.16550196543333126, + "grad_norm": 2.609480857849121, + "learning_rate": 0.0001889675755495934, + "loss": 0.3702, + "step": 10610 + }, + { + "epoch": 0.16565795220565296, + "grad_norm": 1.3105146884918213, + "learning_rate": 0.00018895717643143862, + "loss": 0.3771, + "step": 10620 + }, + { + "epoch": 0.16581393897797467, + "grad_norm": 2.0520312786102295, + "learning_rate": 0.00018894677731328384, + "loss": 0.398, + "step": 10630 + }, + { + "epoch": 0.16596992575029637, + "grad_norm": 0.06613802909851074, + "learning_rate": 0.00018893637819512906, + "loss": 0.1707, + "step": 10640 + }, + { + "epoch": 0.16612591252261807, + "grad_norm": 7.369375228881836, + "learning_rate": 0.00018892597907697428, + "loss": 0.2691, + "step": 10650 + }, + { + "epoch": 0.16628189929493978, + "grad_norm": 3.305032730102539, + "learning_rate": 0.0001889155799588195, + "loss": 0.296, + "step": 10660 + }, + { + "epoch": 0.1664378860672615, + "grad_norm": 3.6533796787261963, + "learning_rate": 0.00018890518084066474, + "loss": 0.2561, + "step": 10670 + }, + { + "epoch": 0.1665938728395832, + "grad_norm": 0.7857325673103333, + "learning_rate": 0.00018889478172250994, + "loss": 0.2949, + "step": 10680 + }, + { + "epoch": 0.16674985961190492, + "grad_norm": 0.9305738210678101, + "learning_rate": 0.00018888438260435518, + "loss": 0.2419, + "step": 10690 + }, + { + "epoch": 0.16690584638422662, + "grad_norm": 3.1299571990966797, + "learning_rate": 0.00018887398348620038, + "loss": 0.3535, + "step": 10700 + }, + { + "epoch": 0.16706183315654832, + "grad_norm": 1.8140242099761963, + "learning_rate": 0.00018886358436804562, + "loss": 0.1652, + "step": 10710 + }, + { + "epoch": 0.16721781992887003, + "grad_norm": 1.815582036972046, + "learning_rate": 0.00018885318524989082, + "loss": 0.3443, + "step": 10720 + }, + { + "epoch": 0.16737380670119173, + "grad_norm": 5.216224670410156, + "learning_rate": 0.00018884278613173603, + "loss": 0.4041, + "step": 10730 + }, + { + "epoch": 0.16752979347351343, + "grad_norm": 1.8566596508026123, + "learning_rate": 0.00018883238701358125, + "loss": 0.3862, + "step": 10740 + }, + { + "epoch": 0.16768578024583516, + "grad_norm": 0.5089476704597473, + "learning_rate": 0.00018882198789542647, + "loss": 0.3982, + "step": 10750 + }, + { + "epoch": 0.16784176701815687, + "grad_norm": 2.2013261318206787, + "learning_rate": 0.0001888115887772717, + "loss": 0.1228, + "step": 10760 + }, + { + "epoch": 0.16799775379047857, + "grad_norm": 2.9381134510040283, + "learning_rate": 0.0001888011896591169, + "loss": 0.3271, + "step": 10770 + }, + { + "epoch": 0.16815374056280027, + "grad_norm": 0.09646535664796829, + "learning_rate": 0.00018879079054096213, + "loss": 0.2909, + "step": 10780 + }, + { + "epoch": 0.16830972733512198, + "grad_norm": 2.3349268436431885, + "learning_rate": 0.00018878039142280735, + "loss": 0.1786, + "step": 10790 + }, + { + "epoch": 0.16846571410744368, + "grad_norm": 0.5867917537689209, + "learning_rate": 0.00018876999230465257, + "loss": 0.3715, + "step": 10800 + }, + { + "epoch": 0.16862170087976538, + "grad_norm": 3.365086793899536, + "learning_rate": 0.0001887595931864978, + "loss": 0.298, + "step": 10810 + }, + { + "epoch": 0.16877768765208712, + "grad_norm": 2.2433769702911377, + "learning_rate": 0.000188749194068343, + "loss": 0.3029, + "step": 10820 + }, + { + "epoch": 0.16893367442440882, + "grad_norm": 3.5059666633605957, + "learning_rate": 0.00018873879495018823, + "loss": 0.2853, + "step": 10830 + }, + { + "epoch": 0.16908966119673052, + "grad_norm": 1.584280252456665, + "learning_rate": 0.00018872839583203345, + "loss": 0.1737, + "step": 10840 + }, + { + "epoch": 0.16924564796905223, + "grad_norm": 2.506237506866455, + "learning_rate": 0.00018871799671387867, + "loss": 0.3987, + "step": 10850 + }, + { + "epoch": 0.16940163474137393, + "grad_norm": 0.34235879778862, + "learning_rate": 0.0001887075975957239, + "loss": 0.1469, + "step": 10860 + }, + { + "epoch": 0.16955762151369563, + "grad_norm": 6.578113555908203, + "learning_rate": 0.0001886971984775691, + "loss": 0.4204, + "step": 10870 + }, + { + "epoch": 0.16971360828601734, + "grad_norm": 1.5973520278930664, + "learning_rate": 0.00018868679935941433, + "loss": 0.3714, + "step": 10880 + }, + { + "epoch": 0.16986959505833904, + "grad_norm": 1.587073802947998, + "learning_rate": 0.00018867640024125955, + "loss": 0.4175, + "step": 10890 + }, + { + "epoch": 0.17002558183066077, + "grad_norm": 2.62223482131958, + "learning_rate": 0.00018866600112310477, + "loss": 0.4588, + "step": 10900 + }, + { + "epoch": 0.17018156860298247, + "grad_norm": 1.3709934949874878, + "learning_rate": 0.00018865560200495, + "loss": 0.3004, + "step": 10910 + }, + { + "epoch": 0.17033755537530418, + "grad_norm": 1.5883265733718872, + "learning_rate": 0.0001886452028867952, + "loss": 0.1858, + "step": 10920 + }, + { + "epoch": 0.17049354214762588, + "grad_norm": 1.151623010635376, + "learning_rate": 0.00018863480376864043, + "loss": 0.4646, + "step": 10930 + }, + { + "epoch": 0.17064952891994759, + "grad_norm": 0.9130459427833557, + "learning_rate": 0.00018862440465048565, + "loss": 0.1834, + "step": 10940 + }, + { + "epoch": 0.1708055156922693, + "grad_norm": 2.551058292388916, + "learning_rate": 0.00018861400553233087, + "loss": 0.3655, + "step": 10950 + }, + { + "epoch": 0.170961502464591, + "grad_norm": 1.753304362297058, + "learning_rate": 0.0001886036064141761, + "loss": 0.2687, + "step": 10960 + }, + { + "epoch": 0.17111748923691272, + "grad_norm": 0.6991026401519775, + "learning_rate": 0.0001885932072960213, + "loss": 0.197, + "step": 10970 + }, + { + "epoch": 0.17127347600923443, + "grad_norm": 2.3284380435943604, + "learning_rate": 0.00018858280817786653, + "loss": 0.2089, + "step": 10980 + }, + { + "epoch": 0.17142946278155613, + "grad_norm": 6.7722649574279785, + "learning_rate": 0.00018857240905971175, + "loss": 0.4157, + "step": 10990 + }, + { + "epoch": 0.17158544955387783, + "grad_norm": 1.0205403566360474, + "learning_rate": 0.00018856200994155697, + "loss": 0.2136, + "step": 11000 + }, + { + "epoch": 0.17174143632619954, + "grad_norm": 3.567028284072876, + "learning_rate": 0.00018855161082340218, + "loss": 0.3153, + "step": 11010 + }, + { + "epoch": 0.17189742309852124, + "grad_norm": 3.081836223602295, + "learning_rate": 0.0001885412117052474, + "loss": 0.3636, + "step": 11020 + }, + { + "epoch": 0.17205340987084294, + "grad_norm": 0.556101381778717, + "learning_rate": 0.00018853081258709262, + "loss": 0.2506, + "step": 11030 + }, + { + "epoch": 0.17220939664316465, + "grad_norm": 2.3356871604919434, + "learning_rate": 0.00018852041346893784, + "loss": 0.3308, + "step": 11040 + }, + { + "epoch": 0.17236538341548638, + "grad_norm": 3.5223493576049805, + "learning_rate": 0.00018851001435078306, + "loss": 0.2867, + "step": 11050 + }, + { + "epoch": 0.17252137018780808, + "grad_norm": 1.1253432035446167, + "learning_rate": 0.00018849961523262828, + "loss": 0.1738, + "step": 11060 + }, + { + "epoch": 0.17267735696012979, + "grad_norm": 1.175328016281128, + "learning_rate": 0.0001884892161144735, + "loss": 0.2748, + "step": 11070 + }, + { + "epoch": 0.1728333437324515, + "grad_norm": 0.33697769045829773, + "learning_rate": 0.00018847881699631872, + "loss": 0.2324, + "step": 11080 + }, + { + "epoch": 0.1729893305047732, + "grad_norm": 3.0209105014801025, + "learning_rate": 0.00018846841787816394, + "loss": 0.1343, + "step": 11090 + }, + { + "epoch": 0.1731453172770949, + "grad_norm": 2.680307388305664, + "learning_rate": 0.00018845801876000916, + "loss": 0.2347, + "step": 11100 + }, + { + "epoch": 0.1733013040494166, + "grad_norm": 4.107524394989014, + "learning_rate": 0.00018844761964185438, + "loss": 0.4449, + "step": 11110 + }, + { + "epoch": 0.17345729082173833, + "grad_norm": 2.8336267471313477, + "learning_rate": 0.0001884372205236996, + "loss": 0.199, + "step": 11120 + }, + { + "epoch": 0.17361327759406003, + "grad_norm": 5.274866104125977, + "learning_rate": 0.00018842682140554482, + "loss": 0.2599, + "step": 11130 + }, + { + "epoch": 0.17376926436638174, + "grad_norm": 0.8743075728416443, + "learning_rate": 0.00018841642228739004, + "loss": 0.2575, + "step": 11140 + }, + { + "epoch": 0.17392525113870344, + "grad_norm": 1.2630996704101562, + "learning_rate": 0.00018840602316923526, + "loss": 0.2234, + "step": 11150 + }, + { + "epoch": 0.17408123791102514, + "grad_norm": 2.6176912784576416, + "learning_rate": 0.00018839562405108048, + "loss": 0.4257, + "step": 11160 + }, + { + "epoch": 0.17423722468334685, + "grad_norm": 4.480280876159668, + "learning_rate": 0.0001883852249329257, + "loss": 0.2601, + "step": 11170 + }, + { + "epoch": 0.17439321145566855, + "grad_norm": 1.2513129711151123, + "learning_rate": 0.00018837482581477092, + "loss": 0.1531, + "step": 11180 + }, + { + "epoch": 0.17454919822799025, + "grad_norm": 4.006619930267334, + "learning_rate": 0.00018836442669661614, + "loss": 0.3433, + "step": 11190 + }, + { + "epoch": 0.17470518500031199, + "grad_norm": 1.13762366771698, + "learning_rate": 0.00018835402757846136, + "loss": 0.322, + "step": 11200 + }, + { + "epoch": 0.1748611717726337, + "grad_norm": 2.1604270935058594, + "learning_rate": 0.00018834362846030658, + "loss": 0.2989, + "step": 11210 + }, + { + "epoch": 0.1750171585449554, + "grad_norm": 1.215661644935608, + "learning_rate": 0.0001883332293421518, + "loss": 0.1577, + "step": 11220 + }, + { + "epoch": 0.1751731453172771, + "grad_norm": 1.3835750818252563, + "learning_rate": 0.00018832283022399702, + "loss": 0.4492, + "step": 11230 + }, + { + "epoch": 0.1753291320895988, + "grad_norm": 2.7984955310821533, + "learning_rate": 0.00018831243110584224, + "loss": 0.4362, + "step": 11240 + }, + { + "epoch": 0.1754851188619205, + "grad_norm": 1.2720012664794922, + "learning_rate": 0.00018830203198768746, + "loss": 0.3137, + "step": 11250 + }, + { + "epoch": 0.1756411056342422, + "grad_norm": 0.7442195415496826, + "learning_rate": 0.00018829163286953268, + "loss": 0.1539, + "step": 11260 + }, + { + "epoch": 0.1757970924065639, + "grad_norm": 2.8686635494232178, + "learning_rate": 0.0001882812337513779, + "loss": 0.3796, + "step": 11270 + }, + { + "epoch": 0.17595307917888564, + "grad_norm": 4.7904815673828125, + "learning_rate": 0.00018827083463322311, + "loss": 0.2852, + "step": 11280 + }, + { + "epoch": 0.17610906595120734, + "grad_norm": 2.434587240219116, + "learning_rate": 0.00018826043551506833, + "loss": 0.3303, + "step": 11290 + }, + { + "epoch": 0.17626505272352905, + "grad_norm": 1.660892128944397, + "learning_rate": 0.00018825003639691355, + "loss": 0.3054, + "step": 11300 + }, + { + "epoch": 0.17642103949585075, + "grad_norm": 5.0868682861328125, + "learning_rate": 0.00018823963727875877, + "loss": 0.3482, + "step": 11310 + }, + { + "epoch": 0.17657702626817245, + "grad_norm": 3.2614431381225586, + "learning_rate": 0.000188229238160604, + "loss": 0.3559, + "step": 11320 + }, + { + "epoch": 0.17673301304049416, + "grad_norm": 2.692411422729492, + "learning_rate": 0.0001882188390424492, + "loss": 0.3964, + "step": 11330 + }, + { + "epoch": 0.17688899981281586, + "grad_norm": 1.3929377794265747, + "learning_rate": 0.00018820843992429443, + "loss": 0.1333, + "step": 11340 + }, + { + "epoch": 0.1770449865851376, + "grad_norm": 0.9341831207275391, + "learning_rate": 0.00018819804080613965, + "loss": 0.2454, + "step": 11350 + }, + { + "epoch": 0.1772009733574593, + "grad_norm": 6.067534446716309, + "learning_rate": 0.00018818764168798487, + "loss": 0.1486, + "step": 11360 + }, + { + "epoch": 0.177356960129781, + "grad_norm": 3.995746612548828, + "learning_rate": 0.0001881772425698301, + "loss": 0.2191, + "step": 11370 + }, + { + "epoch": 0.1775129469021027, + "grad_norm": 0.6745499968528748, + "learning_rate": 0.0001881668434516753, + "loss": 0.2169, + "step": 11380 + }, + { + "epoch": 0.1776689336744244, + "grad_norm": 0.9113379716873169, + "learning_rate": 0.00018815644433352053, + "loss": 0.4376, + "step": 11390 + }, + { + "epoch": 0.1778249204467461, + "grad_norm": 2.4003946781158447, + "learning_rate": 0.00018814604521536575, + "loss": 0.2247, + "step": 11400 + }, + { + "epoch": 0.1779809072190678, + "grad_norm": 0.6229470372200012, + "learning_rate": 0.00018813564609721097, + "loss": 0.3873, + "step": 11410 + }, + { + "epoch": 0.17813689399138952, + "grad_norm": 3.3478331565856934, + "learning_rate": 0.0001881252469790562, + "loss": 0.3678, + "step": 11420 + }, + { + "epoch": 0.17829288076371125, + "grad_norm": 3.337515115737915, + "learning_rate": 0.0001881148478609014, + "loss": 0.257, + "step": 11430 + }, + { + "epoch": 0.17844886753603295, + "grad_norm": 1.0658881664276123, + "learning_rate": 0.00018810444874274663, + "loss": 0.2247, + "step": 11440 + }, + { + "epoch": 0.17860485430835465, + "grad_norm": 1.2788513898849487, + "learning_rate": 0.00018809404962459185, + "loss": 0.155, + "step": 11450 + }, + { + "epoch": 0.17876084108067636, + "grad_norm": 4.128309726715088, + "learning_rate": 0.00018808365050643707, + "loss": 0.4609, + "step": 11460 + }, + { + "epoch": 0.17891682785299806, + "grad_norm": 1.2573262453079224, + "learning_rate": 0.0001880732513882823, + "loss": 0.2594, + "step": 11470 + }, + { + "epoch": 0.17907281462531977, + "grad_norm": 0.6664438843727112, + "learning_rate": 0.0001880628522701275, + "loss": 0.1492, + "step": 11480 + }, + { + "epoch": 0.17922880139764147, + "grad_norm": 1.7848531007766724, + "learning_rate": 0.00018805245315197273, + "loss": 0.3908, + "step": 11490 + }, + { + "epoch": 0.1793847881699632, + "grad_norm": 3.4360225200653076, + "learning_rate": 0.00018804205403381795, + "loss": 0.5462, + "step": 11500 + }, + { + "epoch": 0.1795407749422849, + "grad_norm": 0.7374251484870911, + "learning_rate": 0.00018803165491566317, + "loss": 0.2537, + "step": 11510 + }, + { + "epoch": 0.1796967617146066, + "grad_norm": 1.9689782857894897, + "learning_rate": 0.00018802125579750839, + "loss": 0.3071, + "step": 11520 + }, + { + "epoch": 0.1798527484869283, + "grad_norm": 4.35293436050415, + "learning_rate": 0.0001880108566793536, + "loss": 0.2955, + "step": 11530 + }, + { + "epoch": 0.18000873525925, + "grad_norm": 5.410284519195557, + "learning_rate": 0.00018800045756119883, + "loss": 0.2007, + "step": 11540 + }, + { + "epoch": 0.18016472203157172, + "grad_norm": 0.6631501317024231, + "learning_rate": 0.00018799005844304405, + "loss": 0.4143, + "step": 11550 + }, + { + "epoch": 0.18032070880389342, + "grad_norm": 1.0068690776824951, + "learning_rate": 0.00018797965932488926, + "loss": 0.1777, + "step": 11560 + }, + { + "epoch": 0.18047669557621512, + "grad_norm": 0.6714175939559937, + "learning_rate": 0.00018796926020673448, + "loss": 0.4107, + "step": 11570 + }, + { + "epoch": 0.18063268234853685, + "grad_norm": 2.233462333679199, + "learning_rate": 0.0001879588610885797, + "loss": 0.3198, + "step": 11580 + }, + { + "epoch": 0.18078866912085856, + "grad_norm": 2.7357115745544434, + "learning_rate": 0.00018794846197042492, + "loss": 0.3642, + "step": 11590 + }, + { + "epoch": 0.18094465589318026, + "grad_norm": 0.16101546585559845, + "learning_rate": 0.00018793806285227014, + "loss": 0.3287, + "step": 11600 + }, + { + "epoch": 0.18110064266550197, + "grad_norm": 1.87134850025177, + "learning_rate": 0.00018792766373411536, + "loss": 0.1571, + "step": 11610 + }, + { + "epoch": 0.18125662943782367, + "grad_norm": 1.1603353023529053, + "learning_rate": 0.00018791726461596058, + "loss": 0.3609, + "step": 11620 + }, + { + "epoch": 0.18141261621014537, + "grad_norm": 0.44800999760627747, + "learning_rate": 0.00018790686549780578, + "loss": 0.2677, + "step": 11630 + }, + { + "epoch": 0.18156860298246708, + "grad_norm": 0.45608317852020264, + "learning_rate": 0.00018789646637965102, + "loss": 0.1975, + "step": 11640 + }, + { + "epoch": 0.1817245897547888, + "grad_norm": 2.955970048904419, + "learning_rate": 0.00018788606726149621, + "loss": 0.4535, + "step": 11650 + }, + { + "epoch": 0.1818805765271105, + "grad_norm": 0.37961629033088684, + "learning_rate": 0.00018787566814334146, + "loss": 0.5116, + "step": 11660 + }, + { + "epoch": 0.1820365632994322, + "grad_norm": 2.642307996749878, + "learning_rate": 0.00018786526902518665, + "loss": 0.2784, + "step": 11670 + }, + { + "epoch": 0.18219255007175392, + "grad_norm": 0.8309847712516785, + "learning_rate": 0.0001878548699070319, + "loss": 0.23, + "step": 11680 + }, + { + "epoch": 0.18234853684407562, + "grad_norm": 0.080808624625206, + "learning_rate": 0.0001878444707888771, + "loss": 0.3631, + "step": 11690 + }, + { + "epoch": 0.18250452361639732, + "grad_norm": 0.9569856524467468, + "learning_rate": 0.00018783407167072234, + "loss": 0.2072, + "step": 11700 + }, + { + "epoch": 0.18266051038871903, + "grad_norm": 2.053797483444214, + "learning_rate": 0.00018782367255256753, + "loss": 0.2924, + "step": 11710 + }, + { + "epoch": 0.18281649716104073, + "grad_norm": 0.6872552037239075, + "learning_rate": 0.00018781327343441278, + "loss": 0.4039, + "step": 11720 + }, + { + "epoch": 0.18297248393336246, + "grad_norm": 1.8522825241088867, + "learning_rate": 0.00018780287431625797, + "loss": 0.3471, + "step": 11730 + }, + { + "epoch": 0.18312847070568417, + "grad_norm": 3.0014431476593018, + "learning_rate": 0.00018779247519810322, + "loss": 0.4743, + "step": 11740 + }, + { + "epoch": 0.18328445747800587, + "grad_norm": 5.5990095138549805, + "learning_rate": 0.0001877820760799484, + "loss": 0.3024, + "step": 11750 + }, + { + "epoch": 0.18344044425032757, + "grad_norm": 2.6067895889282227, + "learning_rate": 0.00018777167696179366, + "loss": 0.4603, + "step": 11760 + }, + { + "epoch": 0.18359643102264928, + "grad_norm": 2.759334087371826, + "learning_rate": 0.00018776127784363885, + "loss": 0.2948, + "step": 11770 + }, + { + "epoch": 0.18375241779497098, + "grad_norm": 0.9370937943458557, + "learning_rate": 0.0001877508787254841, + "loss": 0.3548, + "step": 11780 + }, + { + "epoch": 0.18390840456729268, + "grad_norm": 0.2936267554759979, + "learning_rate": 0.0001877404796073293, + "loss": 0.2449, + "step": 11790 + }, + { + "epoch": 0.1840643913396144, + "grad_norm": 2.230560064315796, + "learning_rate": 0.00018773008048917454, + "loss": 0.2819, + "step": 11800 + }, + { + "epoch": 0.18422037811193612, + "grad_norm": 0.9105195999145508, + "learning_rate": 0.00018771968137101973, + "loss": 0.4598, + "step": 11810 + }, + { + "epoch": 0.18437636488425782, + "grad_norm": 3.3687875270843506, + "learning_rate": 0.00018770928225286498, + "loss": 0.2874, + "step": 11820 + }, + { + "epoch": 0.18453235165657952, + "grad_norm": 3.132526397705078, + "learning_rate": 0.00018769888313471017, + "loss": 0.2096, + "step": 11830 + }, + { + "epoch": 0.18468833842890123, + "grad_norm": 2.588862419128418, + "learning_rate": 0.00018768848401655541, + "loss": 0.2246, + "step": 11840 + }, + { + "epoch": 0.18484432520122293, + "grad_norm": 0.9901078939437866, + "learning_rate": 0.0001876780848984006, + "loss": 0.3331, + "step": 11850 + }, + { + "epoch": 0.18500031197354463, + "grad_norm": 1.5796196460723877, + "learning_rate": 0.00018766768578024585, + "loss": 0.3751, + "step": 11860 + }, + { + "epoch": 0.18515629874586634, + "grad_norm": 1.5140503644943237, + "learning_rate": 0.00018765728666209105, + "loss": 0.3795, + "step": 11870 + }, + { + "epoch": 0.18531228551818807, + "grad_norm": 2.5417346954345703, + "learning_rate": 0.0001876468875439363, + "loss": 0.3184, + "step": 11880 + }, + { + "epoch": 0.18546827229050977, + "grad_norm": 0.9759191870689392, + "learning_rate": 0.00018763648842578149, + "loss": 0.1147, + "step": 11890 + }, + { + "epoch": 0.18562425906283148, + "grad_norm": 1.1694114208221436, + "learning_rate": 0.00018762608930762673, + "loss": 0.2551, + "step": 11900 + }, + { + "epoch": 0.18578024583515318, + "grad_norm": 0.10291661322116852, + "learning_rate": 0.00018761569018947193, + "loss": 0.1442, + "step": 11910 + }, + { + "epoch": 0.18593623260747488, + "grad_norm": 4.740861415863037, + "learning_rate": 0.00018760529107131717, + "loss": 0.3442, + "step": 11920 + }, + { + "epoch": 0.1860922193797966, + "grad_norm": 1.1375826597213745, + "learning_rate": 0.00018759489195316236, + "loss": 0.1655, + "step": 11930 + }, + { + "epoch": 0.1862482061521183, + "grad_norm": 2.3239290714263916, + "learning_rate": 0.0001875844928350076, + "loss": 0.4118, + "step": 11940 + }, + { + "epoch": 0.18640419292444002, + "grad_norm": 1.5354567766189575, + "learning_rate": 0.0001875740937168528, + "loss": 0.2956, + "step": 11950 + }, + { + "epoch": 0.18656017969676172, + "grad_norm": 1.871601939201355, + "learning_rate": 0.00018756369459869805, + "loss": 0.1421, + "step": 11960 + }, + { + "epoch": 0.18671616646908343, + "grad_norm": 1.634887456893921, + "learning_rate": 0.00018755329548054324, + "loss": 0.2769, + "step": 11970 + }, + { + "epoch": 0.18687215324140513, + "grad_norm": 3.6324446201324463, + "learning_rate": 0.0001875428963623885, + "loss": 0.5251, + "step": 11980 + }, + { + "epoch": 0.18702814001372683, + "grad_norm": 2.451237440109253, + "learning_rate": 0.00018753249724423368, + "loss": 0.3201, + "step": 11990 + }, + { + "epoch": 0.18718412678604854, + "grad_norm": 0.368939608335495, + "learning_rate": 0.00018752209812607893, + "loss": 0.317, + "step": 12000 + }, + { + "epoch": 0.18734011355837024, + "grad_norm": 0.4980335831642151, + "learning_rate": 0.00018751169900792415, + "loss": 0.24, + "step": 12010 + }, + { + "epoch": 0.18749610033069194, + "grad_norm": 0.7579603791236877, + "learning_rate": 0.00018750129988976937, + "loss": 0.3206, + "step": 12020 + }, + { + "epoch": 0.18765208710301368, + "grad_norm": 2.4506330490112305, + "learning_rate": 0.0001874909007716146, + "loss": 0.4553, + "step": 12030 + }, + { + "epoch": 0.18780807387533538, + "grad_norm": 2.245502233505249, + "learning_rate": 0.0001874805016534598, + "loss": 0.2139, + "step": 12040 + }, + { + "epoch": 0.18796406064765708, + "grad_norm": 1.6800566911697388, + "learning_rate": 0.00018747010253530503, + "loss": 0.2209, + "step": 12050 + }, + { + "epoch": 0.1881200474199788, + "grad_norm": 14.63359260559082, + "learning_rate": 0.00018745970341715025, + "loss": 0.3266, + "step": 12060 + }, + { + "epoch": 0.1882760341923005, + "grad_norm": 0.5305535197257996, + "learning_rate": 0.00018744930429899547, + "loss": 0.1866, + "step": 12070 + }, + { + "epoch": 0.1884320209646222, + "grad_norm": 2.1569783687591553, + "learning_rate": 0.00018743890518084066, + "loss": 0.2674, + "step": 12080 + }, + { + "epoch": 0.1885880077369439, + "grad_norm": 1.8972543478012085, + "learning_rate": 0.0001874285060626859, + "loss": 0.3271, + "step": 12090 + }, + { + "epoch": 0.18874399450926563, + "grad_norm": 1.4007768630981445, + "learning_rate": 0.0001874181069445311, + "loss": 0.3392, + "step": 12100 + }, + { + "epoch": 0.18889998128158733, + "grad_norm": 0.9715983867645264, + "learning_rate": 0.00018740770782637635, + "loss": 0.2061, + "step": 12110 + }, + { + "epoch": 0.18905596805390903, + "grad_norm": 2.8085765838623047, + "learning_rate": 0.00018739730870822154, + "loss": 0.311, + "step": 12120 + }, + { + "epoch": 0.18921195482623074, + "grad_norm": 3.2722768783569336, + "learning_rate": 0.00018738690959006678, + "loss": 0.3657, + "step": 12130 + }, + { + "epoch": 0.18936794159855244, + "grad_norm": 3.875990152359009, + "learning_rate": 0.00018737651047191198, + "loss": 0.2254, + "step": 12140 + }, + { + "epoch": 0.18952392837087415, + "grad_norm": 0.6779225468635559, + "learning_rate": 0.00018736611135375722, + "loss": 0.268, + "step": 12150 + }, + { + "epoch": 0.18967991514319585, + "grad_norm": 1.9641263484954834, + "learning_rate": 0.00018735571223560242, + "loss": 0.4714, + "step": 12160 + }, + { + "epoch": 0.18983590191551755, + "grad_norm": 0.5900086164474487, + "learning_rate": 0.00018734531311744766, + "loss": 0.4215, + "step": 12170 + }, + { + "epoch": 0.18999188868783928, + "grad_norm": 2.9131414890289307, + "learning_rate": 0.00018733491399929286, + "loss": 0.3081, + "step": 12180 + }, + { + "epoch": 0.190147875460161, + "grad_norm": 0.3678191900253296, + "learning_rate": 0.0001873245148811381, + "loss": 0.3538, + "step": 12190 + }, + { + "epoch": 0.1903038622324827, + "grad_norm": 0.8864480257034302, + "learning_rate": 0.0001873141157629833, + "loss": 0.4137, + "step": 12200 + }, + { + "epoch": 0.1904598490048044, + "grad_norm": 1.3851416110992432, + "learning_rate": 0.00018730371664482854, + "loss": 0.3367, + "step": 12210 + }, + { + "epoch": 0.1906158357771261, + "grad_norm": 3.241757392883301, + "learning_rate": 0.00018729331752667373, + "loss": 0.219, + "step": 12220 + }, + { + "epoch": 0.1907718225494478, + "grad_norm": 0.5330601334571838, + "learning_rate": 0.00018728291840851898, + "loss": 0.2083, + "step": 12230 + }, + { + "epoch": 0.1909278093217695, + "grad_norm": 2.9456329345703125, + "learning_rate": 0.00018727251929036417, + "loss": 0.3054, + "step": 12240 + }, + { + "epoch": 0.1910837960940912, + "grad_norm": 1.8985849618911743, + "learning_rate": 0.00018726212017220942, + "loss": 0.3166, + "step": 12250 + }, + { + "epoch": 0.19123978286641294, + "grad_norm": 1.2999069690704346, + "learning_rate": 0.0001872517210540546, + "loss": 0.3138, + "step": 12260 + }, + { + "epoch": 0.19139576963873464, + "grad_norm": 1.8552626371383667, + "learning_rate": 0.00018724132193589986, + "loss": 0.1441, + "step": 12270 + }, + { + "epoch": 0.19155175641105635, + "grad_norm": 1.5392134189605713, + "learning_rate": 0.00018723092281774505, + "loss": 0.2668, + "step": 12280 + }, + { + "epoch": 0.19170774318337805, + "grad_norm": 1.5052801370620728, + "learning_rate": 0.0001872205236995903, + "loss": 0.2145, + "step": 12290 + }, + { + "epoch": 0.19186372995569975, + "grad_norm": 2.863943099975586, + "learning_rate": 0.0001872101245814355, + "loss": 0.2946, + "step": 12300 + }, + { + "epoch": 0.19201971672802146, + "grad_norm": 0.2824110984802246, + "learning_rate": 0.00018719972546328074, + "loss": 0.2881, + "step": 12310 + }, + { + "epoch": 0.19217570350034316, + "grad_norm": 4.316451072692871, + "learning_rate": 0.00018718932634512593, + "loss": 0.1981, + "step": 12320 + }, + { + "epoch": 0.1923316902726649, + "grad_norm": 1.36631441116333, + "learning_rate": 0.00018717892722697118, + "loss": 0.3839, + "step": 12330 + }, + { + "epoch": 0.1924876770449866, + "grad_norm": 1.6842166185379028, + "learning_rate": 0.00018716852810881637, + "loss": 0.1877, + "step": 12340 + }, + { + "epoch": 0.1926436638173083, + "grad_norm": 2.7720024585723877, + "learning_rate": 0.00018715812899066162, + "loss": 0.219, + "step": 12350 + }, + { + "epoch": 0.19279965058963, + "grad_norm": 2.5048646926879883, + "learning_rate": 0.0001871477298725068, + "loss": 0.378, + "step": 12360 + }, + { + "epoch": 0.1929556373619517, + "grad_norm": 1.2577403783798218, + "learning_rate": 0.00018713733075435206, + "loss": 0.2123, + "step": 12370 + }, + { + "epoch": 0.1931116241342734, + "grad_norm": 1.7071456909179688, + "learning_rate": 0.00018712693163619725, + "loss": 0.1933, + "step": 12380 + }, + { + "epoch": 0.1932676109065951, + "grad_norm": 2.5882728099823, + "learning_rate": 0.0001871165325180425, + "loss": 0.2493, + "step": 12390 + }, + { + "epoch": 0.19342359767891681, + "grad_norm": 3.0116987228393555, + "learning_rate": 0.0001871061333998877, + "loss": 0.4508, + "step": 12400 + }, + { + "epoch": 0.19357958445123855, + "grad_norm": 2.4594271183013916, + "learning_rate": 0.00018709573428173293, + "loss": 0.2687, + "step": 12410 + }, + { + "epoch": 0.19373557122356025, + "grad_norm": 1.450259804725647, + "learning_rate": 0.00018708533516357813, + "loss": 0.2859, + "step": 12420 + }, + { + "epoch": 0.19389155799588195, + "grad_norm": 1.4073760509490967, + "learning_rate": 0.00018707493604542337, + "loss": 0.4245, + "step": 12430 + }, + { + "epoch": 0.19404754476820366, + "grad_norm": 2.7500340938568115, + "learning_rate": 0.00018706453692726857, + "loss": 0.3544, + "step": 12440 + }, + { + "epoch": 0.19420353154052536, + "grad_norm": 3.948373556137085, + "learning_rate": 0.0001870541378091138, + "loss": 0.4305, + "step": 12450 + }, + { + "epoch": 0.19435951831284706, + "grad_norm": 0.6741043925285339, + "learning_rate": 0.000187043738690959, + "loss": 0.1207, + "step": 12460 + }, + { + "epoch": 0.19451550508516877, + "grad_norm": 2.580728769302368, + "learning_rate": 0.00018703333957280425, + "loss": 0.4065, + "step": 12470 + }, + { + "epoch": 0.1946714918574905, + "grad_norm": 2.4462509155273438, + "learning_rate": 0.00018702294045464944, + "loss": 0.3055, + "step": 12480 + }, + { + "epoch": 0.1948274786298122, + "grad_norm": 1.164630651473999, + "learning_rate": 0.0001870125413364947, + "loss": 0.2245, + "step": 12490 + }, + { + "epoch": 0.1949834654021339, + "grad_norm": 2.304769515991211, + "learning_rate": 0.00018700214221833988, + "loss": 0.3715, + "step": 12500 + }, + { + "epoch": 0.1951394521744556, + "grad_norm": 1.8169375658035278, + "learning_rate": 0.00018699174310018513, + "loss": 0.387, + "step": 12510 + }, + { + "epoch": 0.1952954389467773, + "grad_norm": 2.0623719692230225, + "learning_rate": 0.00018698134398203032, + "loss": 0.3319, + "step": 12520 + }, + { + "epoch": 0.19545142571909901, + "grad_norm": 1.653314471244812, + "learning_rate": 0.00018697094486387557, + "loss": 0.2448, + "step": 12530 + }, + { + "epoch": 0.19560741249142072, + "grad_norm": 3.3568432331085205, + "learning_rate": 0.00018696054574572076, + "loss": 0.2177, + "step": 12540 + }, + { + "epoch": 0.19576339926374242, + "grad_norm": 1.3817789554595947, + "learning_rate": 0.00018695014662756598, + "loss": 0.1851, + "step": 12550 + }, + { + "epoch": 0.19591938603606415, + "grad_norm": 2.257049083709717, + "learning_rate": 0.0001869397475094112, + "loss": 0.16, + "step": 12560 + }, + { + "epoch": 0.19607537280838586, + "grad_norm": 1.7546944618225098, + "learning_rate": 0.00018692934839125642, + "loss": 0.3812, + "step": 12570 + }, + { + "epoch": 0.19623135958070756, + "grad_norm": 4.909432888031006, + "learning_rate": 0.00018691894927310164, + "loss": 0.2742, + "step": 12580 + }, + { + "epoch": 0.19638734635302926, + "grad_norm": 0.7321370840072632, + "learning_rate": 0.00018690855015494686, + "loss": 0.2173, + "step": 12590 + }, + { + "epoch": 0.19654333312535097, + "grad_norm": 0.7765074968338013, + "learning_rate": 0.00018689815103679208, + "loss": 0.0817, + "step": 12600 + }, + { + "epoch": 0.19669931989767267, + "grad_norm": 2.560410261154175, + "learning_rate": 0.0001868877519186373, + "loss": 0.5002, + "step": 12610 + }, + { + "epoch": 0.19685530666999437, + "grad_norm": 0.5044202208518982, + "learning_rate": 0.00018687735280048252, + "loss": 0.2647, + "step": 12620 + }, + { + "epoch": 0.1970112934423161, + "grad_norm": 6.360370635986328, + "learning_rate": 0.00018686695368232774, + "loss": 0.2416, + "step": 12630 + }, + { + "epoch": 0.1971672802146378, + "grad_norm": 0.19304554164409637, + "learning_rate": 0.00018685655456417296, + "loss": 0.2159, + "step": 12640 + }, + { + "epoch": 0.1973232669869595, + "grad_norm": 1.0849066972732544, + "learning_rate": 0.00018684615544601818, + "loss": 0.2062, + "step": 12650 + }, + { + "epoch": 0.19747925375928121, + "grad_norm": 0.17392005026340485, + "learning_rate": 0.0001868357563278634, + "loss": 0.1824, + "step": 12660 + }, + { + "epoch": 0.19763524053160292, + "grad_norm": 1.6661134958267212, + "learning_rate": 0.00018682535720970862, + "loss": 0.4058, + "step": 12670 + }, + { + "epoch": 0.19779122730392462, + "grad_norm": 2.391458511352539, + "learning_rate": 0.00018681495809155384, + "loss": 0.1612, + "step": 12680 + }, + { + "epoch": 0.19794721407624633, + "grad_norm": 6.055952072143555, + "learning_rate": 0.00018680455897339906, + "loss": 0.18, + "step": 12690 + }, + { + "epoch": 0.19810320084856803, + "grad_norm": 0.36064431071281433, + "learning_rate": 0.00018679415985524428, + "loss": 0.4486, + "step": 12700 + }, + { + "epoch": 0.19825918762088976, + "grad_norm": 3.6982522010803223, + "learning_rate": 0.0001867837607370895, + "loss": 0.268, + "step": 12710 + }, + { + "epoch": 0.19841517439321146, + "grad_norm": 0.039661530405282974, + "learning_rate": 0.00018677336161893472, + "loss": 0.1663, + "step": 12720 + }, + { + "epoch": 0.19857116116553317, + "grad_norm": 6.476260185241699, + "learning_rate": 0.00018676296250077994, + "loss": 0.3868, + "step": 12730 + }, + { + "epoch": 0.19872714793785487, + "grad_norm": 2.246732234954834, + "learning_rate": 0.00018675256338262516, + "loss": 0.3023, + "step": 12740 + }, + { + "epoch": 0.19888313471017657, + "grad_norm": 0.6637864112854004, + "learning_rate": 0.00018674216426447038, + "loss": 0.327, + "step": 12750 + }, + { + "epoch": 0.19903912148249828, + "grad_norm": 7.164607524871826, + "learning_rate": 0.0001867317651463156, + "loss": 0.3744, + "step": 12760 + }, + { + "epoch": 0.19919510825481998, + "grad_norm": 9.32862663269043, + "learning_rate": 0.00018672136602816081, + "loss": 0.3755, + "step": 12770 + }, + { + "epoch": 0.1993510950271417, + "grad_norm": 0.7830541729927063, + "learning_rate": 0.00018671096691000603, + "loss": 0.2464, + "step": 12780 + }, + { + "epoch": 0.19950708179946341, + "grad_norm": 0.5224286913871765, + "learning_rate": 0.00018670056779185125, + "loss": 0.2255, + "step": 12790 + }, + { + "epoch": 0.19966306857178512, + "grad_norm": 2.550424337387085, + "learning_rate": 0.00018669016867369647, + "loss": 0.375, + "step": 12800 + }, + { + "epoch": 0.19981905534410682, + "grad_norm": 2.3978285789489746, + "learning_rate": 0.0001866797695555417, + "loss": 0.133, + "step": 12810 + }, + { + "epoch": 0.19997504211642853, + "grad_norm": 2.235422372817993, + "learning_rate": 0.0001866693704373869, + "loss": 0.2634, + "step": 12820 + }, + { + "epoch": 0.20013102888875023, + "grad_norm": 0.7646775841712952, + "learning_rate": 0.00018665897131923213, + "loss": 0.2511, + "step": 12830 + }, + { + "epoch": 0.20028701566107193, + "grad_norm": 1.183674693107605, + "learning_rate": 0.00018664857220107735, + "loss": 0.4336, + "step": 12840 + }, + { + "epoch": 0.20044300243339364, + "grad_norm": 3.3373656272888184, + "learning_rate": 0.00018663817308292257, + "loss": 0.2061, + "step": 12850 + }, + { + "epoch": 0.20059898920571537, + "grad_norm": 2.4451565742492676, + "learning_rate": 0.00018662777396476782, + "loss": 0.2519, + "step": 12860 + }, + { + "epoch": 0.20075497597803707, + "grad_norm": 2.2898619174957275, + "learning_rate": 0.000186617374846613, + "loss": 0.4251, + "step": 12870 + }, + { + "epoch": 0.20091096275035877, + "grad_norm": 1.5897321701049805, + "learning_rate": 0.00018660697572845826, + "loss": 0.3931, + "step": 12880 + }, + { + "epoch": 0.20106694952268048, + "grad_norm": 0.2315976917743683, + "learning_rate": 0.00018659657661030345, + "loss": 0.2843, + "step": 12890 + }, + { + "epoch": 0.20122293629500218, + "grad_norm": 3.4310429096221924, + "learning_rate": 0.0001865861774921487, + "loss": 0.3827, + "step": 12900 + }, + { + "epoch": 0.20137892306732388, + "grad_norm": 0.6584334373474121, + "learning_rate": 0.0001865757783739939, + "loss": 0.426, + "step": 12910 + }, + { + "epoch": 0.2015349098396456, + "grad_norm": 1.0834167003631592, + "learning_rate": 0.00018656537925583914, + "loss": 0.2333, + "step": 12920 + }, + { + "epoch": 0.20169089661196732, + "grad_norm": 0.3271002769470215, + "learning_rate": 0.00018655498013768433, + "loss": 0.1875, + "step": 12930 + }, + { + "epoch": 0.20184688338428902, + "grad_norm": 0.2775808274745941, + "learning_rate": 0.00018654458101952958, + "loss": 0.2842, + "step": 12940 + }, + { + "epoch": 0.20200287015661073, + "grad_norm": 3.3535640239715576, + "learning_rate": 0.00018653418190137477, + "loss": 0.4467, + "step": 12950 + }, + { + "epoch": 0.20215885692893243, + "grad_norm": 1.405012845993042, + "learning_rate": 0.00018652378278322001, + "loss": 0.3265, + "step": 12960 + }, + { + "epoch": 0.20231484370125413, + "grad_norm": 0.523131251335144, + "learning_rate": 0.0001865133836650652, + "loss": 0.2644, + "step": 12970 + }, + { + "epoch": 0.20247083047357584, + "grad_norm": 2.2922463417053223, + "learning_rate": 0.00018650298454691045, + "loss": 0.4115, + "step": 12980 + }, + { + "epoch": 0.20262681724589754, + "grad_norm": 0.8483665585517883, + "learning_rate": 0.00018649258542875565, + "loss": 0.2821, + "step": 12990 + }, + { + "epoch": 0.20278280401821924, + "grad_norm": 4.608166694641113, + "learning_rate": 0.00018648218631060087, + "loss": 0.4619, + "step": 13000 + }, + { + "epoch": 0.20293879079054097, + "grad_norm": 5.381394863128662, + "learning_rate": 0.00018647178719244609, + "loss": 0.2823, + "step": 13010 + }, + { + "epoch": 0.20309477756286268, + "grad_norm": 0.4596588611602783, + "learning_rate": 0.0001864613880742913, + "loss": 0.4868, + "step": 13020 + }, + { + "epoch": 0.20325076433518438, + "grad_norm": 0.6998667120933533, + "learning_rate": 0.00018645098895613652, + "loss": 0.2351, + "step": 13030 + }, + { + "epoch": 0.20340675110750608, + "grad_norm": 1.0148861408233643, + "learning_rate": 0.00018644058983798174, + "loss": 0.3832, + "step": 13040 + }, + { + "epoch": 0.2035627378798278, + "grad_norm": 1.9544346332550049, + "learning_rate": 0.00018643019071982696, + "loss": 0.2201, + "step": 13050 + }, + { + "epoch": 0.2037187246521495, + "grad_norm": 2.578815460205078, + "learning_rate": 0.00018641979160167218, + "loss": 0.3551, + "step": 13060 + }, + { + "epoch": 0.2038747114244712, + "grad_norm": 1.0851473808288574, + "learning_rate": 0.0001864093924835174, + "loss": 0.318, + "step": 13070 + }, + { + "epoch": 0.20403069819679293, + "grad_norm": 0.012352113611996174, + "learning_rate": 0.00018639899336536262, + "loss": 0.1624, + "step": 13080 + }, + { + "epoch": 0.20418668496911463, + "grad_norm": 4.314785957336426, + "learning_rate": 0.00018638859424720784, + "loss": 0.3751, + "step": 13090 + }, + { + "epoch": 0.20434267174143633, + "grad_norm": 0.8892815709114075, + "learning_rate": 0.00018637819512905306, + "loss": 0.2915, + "step": 13100 + }, + { + "epoch": 0.20449865851375804, + "grad_norm": 0.051064297556877136, + "learning_rate": 0.00018636779601089828, + "loss": 0.2273, + "step": 13110 + }, + { + "epoch": 0.20465464528607974, + "grad_norm": 3.224510908126831, + "learning_rate": 0.0001863573968927435, + "loss": 0.3094, + "step": 13120 + }, + { + "epoch": 0.20481063205840144, + "grad_norm": 2.0588228702545166, + "learning_rate": 0.00018634699777458872, + "loss": 0.3566, + "step": 13130 + }, + { + "epoch": 0.20496661883072315, + "grad_norm": 1.1418896913528442, + "learning_rate": 0.00018633659865643394, + "loss": 0.3485, + "step": 13140 + }, + { + "epoch": 0.20512260560304485, + "grad_norm": 2.586925745010376, + "learning_rate": 0.00018632619953827916, + "loss": 0.3205, + "step": 13150 + }, + { + "epoch": 0.20527859237536658, + "grad_norm": 0.9093180298805237, + "learning_rate": 0.00018631580042012438, + "loss": 0.292, + "step": 13160 + }, + { + "epoch": 0.20543457914768828, + "grad_norm": 1.9040802717208862, + "learning_rate": 0.0001863054013019696, + "loss": 0.2381, + "step": 13170 + }, + { + "epoch": 0.20559056592001, + "grad_norm": 3.7767133712768555, + "learning_rate": 0.00018629500218381482, + "loss": 0.2701, + "step": 13180 + }, + { + "epoch": 0.2057465526923317, + "grad_norm": 0.705057680606842, + "learning_rate": 0.00018628460306566004, + "loss": 0.2575, + "step": 13190 + }, + { + "epoch": 0.2059025394646534, + "grad_norm": 1.85979425907135, + "learning_rate": 0.00018627420394750526, + "loss": 0.3454, + "step": 13200 + }, + { + "epoch": 0.2060585262369751, + "grad_norm": 1.156506896018982, + "learning_rate": 0.00018626380482935048, + "loss": 0.362, + "step": 13210 + }, + { + "epoch": 0.2062145130092968, + "grad_norm": 2.3099493980407715, + "learning_rate": 0.0001862534057111957, + "loss": 0.1441, + "step": 13220 + }, + { + "epoch": 0.2063704997816185, + "grad_norm": 2.2436041831970215, + "learning_rate": 0.00018624300659304092, + "loss": 0.4112, + "step": 13230 + }, + { + "epoch": 0.20652648655394024, + "grad_norm": 1.0545529127120972, + "learning_rate": 0.00018623260747488614, + "loss": 0.1892, + "step": 13240 + }, + { + "epoch": 0.20668247332626194, + "grad_norm": 0.11491916328668594, + "learning_rate": 0.00018622220835673136, + "loss": 0.096, + "step": 13250 + }, + { + "epoch": 0.20683846009858364, + "grad_norm": 20.04274559020996, + "learning_rate": 0.00018621180923857658, + "loss": 0.5643, + "step": 13260 + }, + { + "epoch": 0.20699444687090535, + "grad_norm": 0.3301975429058075, + "learning_rate": 0.0001862014101204218, + "loss": 0.1799, + "step": 13270 + }, + { + "epoch": 0.20715043364322705, + "grad_norm": 1.0274220705032349, + "learning_rate": 0.00018619101100226702, + "loss": 0.22, + "step": 13280 + }, + { + "epoch": 0.20730642041554875, + "grad_norm": 0.966164231300354, + "learning_rate": 0.00018618061188411224, + "loss": 0.2867, + "step": 13290 + }, + { + "epoch": 0.20746240718787046, + "grad_norm": 2.4565162658691406, + "learning_rate": 0.00018617021276595746, + "loss": 0.1691, + "step": 13300 + }, + { + "epoch": 0.2076183939601922, + "grad_norm": 1.225339412689209, + "learning_rate": 0.00018615981364780267, + "loss": 0.301, + "step": 13310 + }, + { + "epoch": 0.2077743807325139, + "grad_norm": 1.0022715330123901, + "learning_rate": 0.0001861494145296479, + "loss": 0.2346, + "step": 13320 + }, + { + "epoch": 0.2079303675048356, + "grad_norm": 1.1647560596466064, + "learning_rate": 0.00018613901541149311, + "loss": 0.1874, + "step": 13330 + }, + { + "epoch": 0.2080863542771573, + "grad_norm": 0.1108563095331192, + "learning_rate": 0.00018612861629333833, + "loss": 0.2054, + "step": 13340 + }, + { + "epoch": 0.208242341049479, + "grad_norm": 1.531417727470398, + "learning_rate": 0.00018611821717518355, + "loss": 0.2334, + "step": 13350 + }, + { + "epoch": 0.2083983278218007, + "grad_norm": 8.367430686950684, + "learning_rate": 0.00018610781805702877, + "loss": 0.4332, + "step": 13360 + }, + { + "epoch": 0.2085543145941224, + "grad_norm": 3.636106491088867, + "learning_rate": 0.000186097418938874, + "loss": 0.3066, + "step": 13370 + }, + { + "epoch": 0.2087103013664441, + "grad_norm": 1.2621756792068481, + "learning_rate": 0.0001860870198207192, + "loss": 0.1554, + "step": 13380 + }, + { + "epoch": 0.20886628813876584, + "grad_norm": 2.553389310836792, + "learning_rate": 0.00018607662070256443, + "loss": 0.3637, + "step": 13390 + }, + { + "epoch": 0.20902227491108755, + "grad_norm": 3.1005523204803467, + "learning_rate": 0.00018606622158440965, + "loss": 0.4428, + "step": 13400 + }, + { + "epoch": 0.20917826168340925, + "grad_norm": 2.580308437347412, + "learning_rate": 0.00018605582246625487, + "loss": 0.3802, + "step": 13410 + }, + { + "epoch": 0.20933424845573095, + "grad_norm": 4.594557762145996, + "learning_rate": 0.0001860454233481001, + "loss": 0.389, + "step": 13420 + }, + { + "epoch": 0.20949023522805266, + "grad_norm": 1.388816475868225, + "learning_rate": 0.0001860350242299453, + "loss": 0.3143, + "step": 13430 + }, + { + "epoch": 0.20964622200037436, + "grad_norm": 1.9355179071426392, + "learning_rate": 0.00018602462511179053, + "loss": 0.2166, + "step": 13440 + }, + { + "epoch": 0.20980220877269606, + "grad_norm": 0.3161306381225586, + "learning_rate": 0.00018601422599363575, + "loss": 0.178, + "step": 13450 + }, + { + "epoch": 0.2099581955450178, + "grad_norm": 3.02839994430542, + "learning_rate": 0.00018600382687548097, + "loss": 0.3922, + "step": 13460 + }, + { + "epoch": 0.2101141823173395, + "grad_norm": 4.012632369995117, + "learning_rate": 0.0001859934277573262, + "loss": 0.2692, + "step": 13470 + }, + { + "epoch": 0.2102701690896612, + "grad_norm": 0.4715295135974884, + "learning_rate": 0.0001859830286391714, + "loss": 0.2473, + "step": 13480 + }, + { + "epoch": 0.2104261558619829, + "grad_norm": 4.312544345855713, + "learning_rate": 0.00018597262952101663, + "loss": 0.3971, + "step": 13490 + }, + { + "epoch": 0.2105821426343046, + "grad_norm": 2.733032703399658, + "learning_rate": 0.00018596223040286185, + "loss": 0.2351, + "step": 13500 + }, + { + "epoch": 0.2107381294066263, + "grad_norm": 0.5007910132408142, + "learning_rate": 0.00018595183128470707, + "loss": 0.2017, + "step": 13510 + }, + { + "epoch": 0.21089411617894802, + "grad_norm": 2.288851499557495, + "learning_rate": 0.0001859414321665523, + "loss": 0.2091, + "step": 13520 + }, + { + "epoch": 0.21105010295126972, + "grad_norm": 0.2666260898113251, + "learning_rate": 0.0001859310330483975, + "loss": 0.151, + "step": 13530 + }, + { + "epoch": 0.21120608972359145, + "grad_norm": 1.1906909942626953, + "learning_rate": 0.00018592063393024273, + "loss": 0.204, + "step": 13540 + }, + { + "epoch": 0.21136207649591315, + "grad_norm": 0.8535648584365845, + "learning_rate": 0.00018591023481208795, + "loss": 0.2528, + "step": 13550 + }, + { + "epoch": 0.21151806326823486, + "grad_norm": 0.27986058592796326, + "learning_rate": 0.00018589983569393317, + "loss": 0.3157, + "step": 13560 + }, + { + "epoch": 0.21167405004055656, + "grad_norm": 2.745215892791748, + "learning_rate": 0.00018588943657577839, + "loss": 0.2119, + "step": 13570 + }, + { + "epoch": 0.21183003681287826, + "grad_norm": 3.1329569816589355, + "learning_rate": 0.0001858790374576236, + "loss": 0.5207, + "step": 13580 + }, + { + "epoch": 0.21198602358519997, + "grad_norm": 3.0692920684814453, + "learning_rate": 0.00018586863833946882, + "loss": 0.4418, + "step": 13590 + }, + { + "epoch": 0.21214201035752167, + "grad_norm": 1.4698386192321777, + "learning_rate": 0.00018585823922131404, + "loss": 0.4403, + "step": 13600 + }, + { + "epoch": 0.2122979971298434, + "grad_norm": 4.66435432434082, + "learning_rate": 0.00018584784010315926, + "loss": 0.3417, + "step": 13610 + }, + { + "epoch": 0.2124539839021651, + "grad_norm": 0.7586390376091003, + "learning_rate": 0.00018583744098500448, + "loss": 0.4965, + "step": 13620 + }, + { + "epoch": 0.2126099706744868, + "grad_norm": 0.4069293737411499, + "learning_rate": 0.0001858270418668497, + "loss": 0.3336, + "step": 13630 + }, + { + "epoch": 0.2127659574468085, + "grad_norm": 1.5166949033737183, + "learning_rate": 0.00018581664274869492, + "loss": 0.3562, + "step": 13640 + }, + { + "epoch": 0.21292194421913022, + "grad_norm": 1.1084874868392944, + "learning_rate": 0.00018580624363054014, + "loss": 0.268, + "step": 13650 + }, + { + "epoch": 0.21307793099145192, + "grad_norm": 0.45353636145591736, + "learning_rate": 0.00018579584451238536, + "loss": 0.098, + "step": 13660 + }, + { + "epoch": 0.21323391776377362, + "grad_norm": 0.04974639415740967, + "learning_rate": 0.00018578544539423058, + "loss": 0.1093, + "step": 13670 + }, + { + "epoch": 0.21338990453609533, + "grad_norm": 1.2779793739318848, + "learning_rate": 0.0001857750462760758, + "loss": 0.1467, + "step": 13680 + }, + { + "epoch": 0.21354589130841706, + "grad_norm": 2.6664750576019287, + "learning_rate": 0.00018576464715792102, + "loss": 0.1693, + "step": 13690 + }, + { + "epoch": 0.21370187808073876, + "grad_norm": 2.0953176021575928, + "learning_rate": 0.00018575424803976624, + "loss": 0.1589, + "step": 13700 + }, + { + "epoch": 0.21385786485306046, + "grad_norm": 3.7887356281280518, + "learning_rate": 0.00018574384892161146, + "loss": 0.2345, + "step": 13710 + }, + { + "epoch": 0.21401385162538217, + "grad_norm": 1.180552363395691, + "learning_rate": 0.00018573344980345668, + "loss": 0.4337, + "step": 13720 + }, + { + "epoch": 0.21416983839770387, + "grad_norm": 2.7489490509033203, + "learning_rate": 0.0001857230506853019, + "loss": 0.4024, + "step": 13730 + }, + { + "epoch": 0.21432582517002557, + "grad_norm": 1.038478970527649, + "learning_rate": 0.00018571265156714712, + "loss": 0.2093, + "step": 13740 + }, + { + "epoch": 0.21448181194234728, + "grad_norm": 0.48595181107521057, + "learning_rate": 0.00018570225244899234, + "loss": 0.1071, + "step": 13750 + }, + { + "epoch": 0.214637798714669, + "grad_norm": 0.2550446093082428, + "learning_rate": 0.00018569185333083756, + "loss": 0.3039, + "step": 13760 + }, + { + "epoch": 0.2147937854869907, + "grad_norm": 1.5095396041870117, + "learning_rate": 0.00018568145421268278, + "loss": 0.2183, + "step": 13770 + }, + { + "epoch": 0.21494977225931242, + "grad_norm": 2.375882148742676, + "learning_rate": 0.000185671055094528, + "loss": 0.2727, + "step": 13780 + }, + { + "epoch": 0.21510575903163412, + "grad_norm": 6.080036163330078, + "learning_rate": 0.00018566065597637322, + "loss": 0.2297, + "step": 13790 + }, + { + "epoch": 0.21526174580395582, + "grad_norm": 3.425128698348999, + "learning_rate": 0.00018565025685821844, + "loss": 0.613, + "step": 13800 + }, + { + "epoch": 0.21541773257627753, + "grad_norm": 0.8501274585723877, + "learning_rate": 0.00018563985774006366, + "loss": 0.3014, + "step": 13810 + }, + { + "epoch": 0.21557371934859923, + "grad_norm": 0.09150723367929459, + "learning_rate": 0.00018562945862190888, + "loss": 0.1303, + "step": 13820 + }, + { + "epoch": 0.21572970612092093, + "grad_norm": 1.3347816467285156, + "learning_rate": 0.0001856190595037541, + "loss": 0.1342, + "step": 13830 + }, + { + "epoch": 0.21588569289324266, + "grad_norm": 3.5113649368286133, + "learning_rate": 0.00018560866038559932, + "loss": 0.5422, + "step": 13840 + }, + { + "epoch": 0.21604167966556437, + "grad_norm": 0.23474463820457458, + "learning_rate": 0.00018559826126744454, + "loss": 0.2288, + "step": 13850 + }, + { + "epoch": 0.21619766643788607, + "grad_norm": 2.839104175567627, + "learning_rate": 0.00018558786214928976, + "loss": 0.219, + "step": 13860 + }, + { + "epoch": 0.21635365321020777, + "grad_norm": 0.34769389033317566, + "learning_rate": 0.00018557746303113497, + "loss": 0.2222, + "step": 13870 + }, + { + "epoch": 0.21650963998252948, + "grad_norm": 1.594574213027954, + "learning_rate": 0.0001855670639129802, + "loss": 0.2995, + "step": 13880 + }, + { + "epoch": 0.21666562675485118, + "grad_norm": 0.2993415594100952, + "learning_rate": 0.00018555666479482541, + "loss": 0.3873, + "step": 13890 + }, + { + "epoch": 0.21682161352717289, + "grad_norm": 1.1278647184371948, + "learning_rate": 0.0001855462656766706, + "loss": 0.2338, + "step": 13900 + }, + { + "epoch": 0.21697760029949462, + "grad_norm": 2.5761172771453857, + "learning_rate": 0.00018553586655851585, + "loss": 0.333, + "step": 13910 + }, + { + "epoch": 0.21713358707181632, + "grad_norm": 0.07127764075994492, + "learning_rate": 0.00018552546744036105, + "loss": 0.3402, + "step": 13920 + }, + { + "epoch": 0.21728957384413802, + "grad_norm": 5.877762317657471, + "learning_rate": 0.0001855150683222063, + "loss": 0.1648, + "step": 13930 + }, + { + "epoch": 0.21744556061645973, + "grad_norm": 0.13584519922733307, + "learning_rate": 0.00018550466920405149, + "loss": 0.1926, + "step": 13940 + }, + { + "epoch": 0.21760154738878143, + "grad_norm": 3.3923139572143555, + "learning_rate": 0.00018549427008589673, + "loss": 0.1384, + "step": 13950 + }, + { + "epoch": 0.21775753416110313, + "grad_norm": 0.06256578862667084, + "learning_rate": 0.00018548387096774192, + "loss": 0.1352, + "step": 13960 + }, + { + "epoch": 0.21791352093342484, + "grad_norm": 1.0103802680969238, + "learning_rate": 0.00018547347184958717, + "loss": 0.0949, + "step": 13970 + }, + { + "epoch": 0.21806950770574654, + "grad_norm": 0.15761008858680725, + "learning_rate": 0.00018546307273143236, + "loss": 0.475, + "step": 13980 + }, + { + "epoch": 0.21822549447806827, + "grad_norm": 0.45524874329566956, + "learning_rate": 0.0001854526736132776, + "loss": 0.2054, + "step": 13990 + }, + { + "epoch": 0.21838148125038997, + "grad_norm": 4.916372776031494, + "learning_rate": 0.0001854422744951228, + "loss": 0.2798, + "step": 14000 + }, + { + "epoch": 0.21853746802271168, + "grad_norm": 6.627929210662842, + "learning_rate": 0.00018543187537696805, + "loss": 0.3078, + "step": 14010 + }, + { + "epoch": 0.21869345479503338, + "grad_norm": 1.1391568183898926, + "learning_rate": 0.00018542147625881324, + "loss": 0.1718, + "step": 14020 + }, + { + "epoch": 0.21884944156735509, + "grad_norm": 0.17036092281341553, + "learning_rate": 0.0001854110771406585, + "loss": 0.3333, + "step": 14030 + }, + { + "epoch": 0.2190054283396768, + "grad_norm": 0.6396644711494446, + "learning_rate": 0.00018540067802250368, + "loss": 0.3808, + "step": 14040 + }, + { + "epoch": 0.2191614151119985, + "grad_norm": 0.9471602439880371, + "learning_rate": 0.00018539027890434893, + "loss": 0.2925, + "step": 14050 + }, + { + "epoch": 0.21931740188432022, + "grad_norm": 2.7182493209838867, + "learning_rate": 0.00018537987978619412, + "loss": 0.2953, + "step": 14060 + }, + { + "epoch": 0.21947338865664193, + "grad_norm": 1.9284687042236328, + "learning_rate": 0.00018536948066803937, + "loss": 0.4335, + "step": 14070 + }, + { + "epoch": 0.21962937542896363, + "grad_norm": 0.6150299906730652, + "learning_rate": 0.00018535908154988456, + "loss": 0.2903, + "step": 14080 + }, + { + "epoch": 0.21978536220128533, + "grad_norm": 1.0829017162322998, + "learning_rate": 0.0001853486824317298, + "loss": 0.1105, + "step": 14090 + }, + { + "epoch": 0.21994134897360704, + "grad_norm": 1.466137409210205, + "learning_rate": 0.000185338283313575, + "loss": 0.287, + "step": 14100 + }, + { + "epoch": 0.22009733574592874, + "grad_norm": 1.9095147848129272, + "learning_rate": 0.00018532788419542025, + "loss": 0.2381, + "step": 14110 + }, + { + "epoch": 0.22025332251825044, + "grad_norm": 1.4633227586746216, + "learning_rate": 0.00018531748507726544, + "loss": 0.6719, + "step": 14120 + }, + { + "epoch": 0.22040930929057215, + "grad_norm": 0.3437176048755646, + "learning_rate": 0.00018530708595911069, + "loss": 0.3679, + "step": 14130 + }, + { + "epoch": 0.22056529606289388, + "grad_norm": 0.44980284571647644, + "learning_rate": 0.00018529668684095588, + "loss": 0.3156, + "step": 14140 + }, + { + "epoch": 0.22072128283521558, + "grad_norm": 6.884535789489746, + "learning_rate": 0.00018528628772280112, + "loss": 0.2477, + "step": 14150 + }, + { + "epoch": 0.22087726960753729, + "grad_norm": 0.6073703765869141, + "learning_rate": 0.00018527588860464632, + "loss": 0.3014, + "step": 14160 + }, + { + "epoch": 0.221033256379859, + "grad_norm": 0.2610427439212799, + "learning_rate": 0.00018526548948649156, + "loss": 0.1413, + "step": 14170 + }, + { + "epoch": 0.2211892431521807, + "grad_norm": 1.1830434799194336, + "learning_rate": 0.00018525509036833676, + "loss": 0.3642, + "step": 14180 + }, + { + "epoch": 0.2213452299245024, + "grad_norm": 0.7711977362632751, + "learning_rate": 0.000185244691250182, + "loss": 0.1781, + "step": 14190 + }, + { + "epoch": 0.2215012166968241, + "grad_norm": 2.457845449447632, + "learning_rate": 0.00018523429213202722, + "loss": 0.4993, + "step": 14200 + }, + { + "epoch": 0.2216572034691458, + "grad_norm": 0.10955695062875748, + "learning_rate": 0.00018522389301387244, + "loss": 0.1925, + "step": 14210 + }, + { + "epoch": 0.22181319024146753, + "grad_norm": 0.2178761214017868, + "learning_rate": 0.00018521349389571766, + "loss": 0.3793, + "step": 14220 + }, + { + "epoch": 0.22196917701378924, + "grad_norm": 0.16451112926006317, + "learning_rate": 0.00018520309477756288, + "loss": 0.429, + "step": 14230 + }, + { + "epoch": 0.22212516378611094, + "grad_norm": 0.6780962347984314, + "learning_rate": 0.0001851926956594081, + "loss": 0.2172, + "step": 14240 + }, + { + "epoch": 0.22228115055843264, + "grad_norm": 0.36148059368133545, + "learning_rate": 0.00018518229654125332, + "loss": 0.2296, + "step": 14250 + }, + { + "epoch": 0.22243713733075435, + "grad_norm": 1.212991714477539, + "learning_rate": 0.00018517189742309854, + "loss": 0.2946, + "step": 14260 + }, + { + "epoch": 0.22259312410307605, + "grad_norm": 5.023010730743408, + "learning_rate": 0.00018516149830494376, + "loss": 0.2245, + "step": 14270 + }, + { + "epoch": 0.22274911087539775, + "grad_norm": 0.5933042168617249, + "learning_rate": 0.00018515109918678898, + "loss": 0.2727, + "step": 14280 + }, + { + "epoch": 0.22290509764771949, + "grad_norm": 0.33793121576309204, + "learning_rate": 0.0001851407000686342, + "loss": 0.1686, + "step": 14290 + }, + { + "epoch": 0.2230610844200412, + "grad_norm": 2.3100786209106445, + "learning_rate": 0.00018513030095047942, + "loss": 0.3075, + "step": 14300 + }, + { + "epoch": 0.2232170711923629, + "grad_norm": 1.0906388759613037, + "learning_rate": 0.00018511990183232464, + "loss": 0.3257, + "step": 14310 + }, + { + "epoch": 0.2233730579646846, + "grad_norm": 3.005523443222046, + "learning_rate": 0.00018510950271416986, + "loss": 0.3126, + "step": 14320 + }, + { + "epoch": 0.2235290447370063, + "grad_norm": 0.589547336101532, + "learning_rate": 0.00018509910359601508, + "loss": 0.218, + "step": 14330 + }, + { + "epoch": 0.223685031509328, + "grad_norm": 2.674639940261841, + "learning_rate": 0.0001850887044778603, + "loss": 0.281, + "step": 14340 + }, + { + "epoch": 0.2238410182816497, + "grad_norm": 0.6713445782661438, + "learning_rate": 0.0001850783053597055, + "loss": 0.4374, + "step": 14350 + }, + { + "epoch": 0.2239970050539714, + "grad_norm": 3.0313546657562256, + "learning_rate": 0.00018506790624155074, + "loss": 0.3541, + "step": 14360 + }, + { + "epoch": 0.22415299182629314, + "grad_norm": 3.273052215576172, + "learning_rate": 0.00018505750712339593, + "loss": 0.1917, + "step": 14370 + }, + { + "epoch": 0.22430897859861484, + "grad_norm": 0.43452969193458557, + "learning_rate": 0.00018504710800524118, + "loss": 0.4404, + "step": 14380 + }, + { + "epoch": 0.22446496537093655, + "grad_norm": 1.0550246238708496, + "learning_rate": 0.00018503670888708637, + "loss": 0.2076, + "step": 14390 + }, + { + "epoch": 0.22462095214325825, + "grad_norm": 0.5780364871025085, + "learning_rate": 0.00018502630976893162, + "loss": 0.1994, + "step": 14400 + }, + { + "epoch": 0.22477693891557995, + "grad_norm": 6.360426902770996, + "learning_rate": 0.0001850159106507768, + "loss": 0.3267, + "step": 14410 + }, + { + "epoch": 0.22493292568790166, + "grad_norm": 3.190255880355835, + "learning_rate": 0.00018500551153262206, + "loss": 0.4444, + "step": 14420 + }, + { + "epoch": 0.22508891246022336, + "grad_norm": 1.1691663265228271, + "learning_rate": 0.00018499511241446725, + "loss": 0.4057, + "step": 14430 + }, + { + "epoch": 0.2252448992325451, + "grad_norm": 0.6865898966789246, + "learning_rate": 0.0001849847132963125, + "loss": 0.3189, + "step": 14440 + }, + { + "epoch": 0.2254008860048668, + "grad_norm": 1.6037561893463135, + "learning_rate": 0.0001849743141781577, + "loss": 0.1874, + "step": 14450 + }, + { + "epoch": 0.2255568727771885, + "grad_norm": 4.867454528808594, + "learning_rate": 0.00018496391506000293, + "loss": 0.5804, + "step": 14460 + }, + { + "epoch": 0.2257128595495102, + "grad_norm": 1.7503598928451538, + "learning_rate": 0.00018495351594184813, + "loss": 0.3266, + "step": 14470 + }, + { + "epoch": 0.2258688463218319, + "grad_norm": 4.39780855178833, + "learning_rate": 0.00018494311682369337, + "loss": 0.3551, + "step": 14480 + }, + { + "epoch": 0.2260248330941536, + "grad_norm": 1.8512024879455566, + "learning_rate": 0.00018493271770553857, + "loss": 0.3033, + "step": 14490 + }, + { + "epoch": 0.2261808198664753, + "grad_norm": 0.6735418438911438, + "learning_rate": 0.0001849223185873838, + "loss": 0.234, + "step": 14500 + }, + { + "epoch": 0.22633680663879702, + "grad_norm": 2.0303027629852295, + "learning_rate": 0.000184911919469229, + "loss": 0.3693, + "step": 14510 + }, + { + "epoch": 0.22649279341111875, + "grad_norm": 1.9029146432876587, + "learning_rate": 0.00018490152035107425, + "loss": 0.3982, + "step": 14520 + }, + { + "epoch": 0.22664878018344045, + "grad_norm": 0.22184059023857117, + "learning_rate": 0.00018489112123291944, + "loss": 0.1491, + "step": 14530 + }, + { + "epoch": 0.22680476695576215, + "grad_norm": 2.114410400390625, + "learning_rate": 0.0001848807221147647, + "loss": 0.353, + "step": 14540 + }, + { + "epoch": 0.22696075372808386, + "grad_norm": 0.8823143839836121, + "learning_rate": 0.00018487032299660988, + "loss": 0.2404, + "step": 14550 + }, + { + "epoch": 0.22711674050040556, + "grad_norm": 6.418842792510986, + "learning_rate": 0.00018485992387845513, + "loss": 0.5292, + "step": 14560 + }, + { + "epoch": 0.22727272727272727, + "grad_norm": 4.001767158508301, + "learning_rate": 0.00018484952476030032, + "loss": 0.3207, + "step": 14570 + }, + { + "epoch": 0.22742871404504897, + "grad_norm": 2.609896659851074, + "learning_rate": 0.00018483912564214557, + "loss": 0.1155, + "step": 14580 + }, + { + "epoch": 0.2275847008173707, + "grad_norm": 5.008525371551514, + "learning_rate": 0.00018482872652399076, + "loss": 0.2064, + "step": 14590 + }, + { + "epoch": 0.2277406875896924, + "grad_norm": 1.322837233543396, + "learning_rate": 0.000184818327405836, + "loss": 0.3166, + "step": 14600 + }, + { + "epoch": 0.2278966743620141, + "grad_norm": 1.3244410753250122, + "learning_rate": 0.0001848079282876812, + "loss": 0.2925, + "step": 14610 + }, + { + "epoch": 0.2280526611343358, + "grad_norm": 5.076612949371338, + "learning_rate": 0.00018479752916952645, + "loss": 0.6025, + "step": 14620 + }, + { + "epoch": 0.2282086479066575, + "grad_norm": 0.476762980222702, + "learning_rate": 0.00018478713005137164, + "loss": 0.1536, + "step": 14630 + }, + { + "epoch": 0.22836463467897922, + "grad_norm": 1.2489818334579468, + "learning_rate": 0.0001847767309332169, + "loss": 0.3104, + "step": 14640 + }, + { + "epoch": 0.22852062145130092, + "grad_norm": 3.2772064208984375, + "learning_rate": 0.00018476633181506208, + "loss": 0.2487, + "step": 14650 + }, + { + "epoch": 0.22867660822362262, + "grad_norm": 8.088106155395508, + "learning_rate": 0.00018475593269690733, + "loss": 0.2436, + "step": 14660 + }, + { + "epoch": 0.22883259499594436, + "grad_norm": 1.7012649774551392, + "learning_rate": 0.00018474553357875252, + "loss": 0.16, + "step": 14670 + }, + { + "epoch": 0.22898858176826606, + "grad_norm": 2.5969321727752686, + "learning_rate": 0.00018473513446059777, + "loss": 0.2961, + "step": 14680 + }, + { + "epoch": 0.22914456854058776, + "grad_norm": 0.9858652353286743, + "learning_rate": 0.00018472473534244296, + "loss": 0.1615, + "step": 14690 + }, + { + "epoch": 0.22930055531290947, + "grad_norm": 2.701960325241089, + "learning_rate": 0.0001847143362242882, + "loss": 0.278, + "step": 14700 + }, + { + "epoch": 0.22945654208523117, + "grad_norm": 0.3224952816963196, + "learning_rate": 0.0001847039371061334, + "loss": 0.1611, + "step": 14710 + }, + { + "epoch": 0.22961252885755287, + "grad_norm": 1.3847914934158325, + "learning_rate": 0.00018469353798797864, + "loss": 0.5682, + "step": 14720 + }, + { + "epoch": 0.22976851562987458, + "grad_norm": 0.5572181940078735, + "learning_rate": 0.00018468313886982384, + "loss": 0.361, + "step": 14730 + }, + { + "epoch": 0.2299245024021963, + "grad_norm": 1.4932482242584229, + "learning_rate": 0.00018467273975166908, + "loss": 0.2427, + "step": 14740 + }, + { + "epoch": 0.230080489174518, + "grad_norm": 5.563808917999268, + "learning_rate": 0.00018466234063351428, + "loss": 0.4409, + "step": 14750 + }, + { + "epoch": 0.2302364759468397, + "grad_norm": 0.9475265145301819, + "learning_rate": 0.00018465194151535952, + "loss": 0.4512, + "step": 14760 + }, + { + "epoch": 0.23039246271916142, + "grad_norm": 1.8160616159439087, + "learning_rate": 0.00018464154239720472, + "loss": 0.306, + "step": 14770 + }, + { + "epoch": 0.23054844949148312, + "grad_norm": 1.0732847452163696, + "learning_rate": 0.00018463114327904996, + "loss": 0.1191, + "step": 14780 + }, + { + "epoch": 0.23070443626380482, + "grad_norm": 2.0108041763305664, + "learning_rate": 0.00018462074416089515, + "loss": 0.3802, + "step": 14790 + }, + { + "epoch": 0.23086042303612653, + "grad_norm": 3.6508185863494873, + "learning_rate": 0.00018461034504274037, + "loss": 0.4014, + "step": 14800 + }, + { + "epoch": 0.23101640980844823, + "grad_norm": 0.8693599700927734, + "learning_rate": 0.0001845999459245856, + "loss": 0.245, + "step": 14810 + }, + { + "epoch": 0.23117239658076996, + "grad_norm": 0.01044541783630848, + "learning_rate": 0.00018458954680643081, + "loss": 0.2118, + "step": 14820 + }, + { + "epoch": 0.23132838335309167, + "grad_norm": 3.852083206176758, + "learning_rate": 0.00018457914768827603, + "loss": 0.2846, + "step": 14830 + }, + { + "epoch": 0.23148437012541337, + "grad_norm": 3.7660000324249268, + "learning_rate": 0.00018456874857012125, + "loss": 0.212, + "step": 14840 + }, + { + "epoch": 0.23164035689773507, + "grad_norm": 3.2255451679229736, + "learning_rate": 0.00018455834945196647, + "loss": 0.4362, + "step": 14850 + }, + { + "epoch": 0.23179634367005678, + "grad_norm": 1.00737464427948, + "learning_rate": 0.0001845479503338117, + "loss": 0.2798, + "step": 14860 + }, + { + "epoch": 0.23195233044237848, + "grad_norm": 1.0015833377838135, + "learning_rate": 0.0001845375512156569, + "loss": 0.313, + "step": 14870 + }, + { + "epoch": 0.23210831721470018, + "grad_norm": 9.4893217086792, + "learning_rate": 0.00018452715209750213, + "loss": 0.3578, + "step": 14880 + }, + { + "epoch": 0.23226430398702191, + "grad_norm": 0.407869815826416, + "learning_rate": 0.00018451675297934735, + "loss": 0.2632, + "step": 14890 + }, + { + "epoch": 0.23242029075934362, + "grad_norm": 1.4509222507476807, + "learning_rate": 0.00018450635386119257, + "loss": 0.4806, + "step": 14900 + }, + { + "epoch": 0.23257627753166532, + "grad_norm": 8.443758964538574, + "learning_rate": 0.0001844959547430378, + "loss": 0.2039, + "step": 14910 + }, + { + "epoch": 0.23273226430398702, + "grad_norm": 2.631842851638794, + "learning_rate": 0.000184485555624883, + "loss": 0.4461, + "step": 14920 + }, + { + "epoch": 0.23288825107630873, + "grad_norm": 3.4924230575561523, + "learning_rate": 0.00018447515650672823, + "loss": 0.2307, + "step": 14930 + }, + { + "epoch": 0.23304423784863043, + "grad_norm": 0.9378137588500977, + "learning_rate": 0.00018446475738857345, + "loss": 0.3553, + "step": 14940 + }, + { + "epoch": 0.23320022462095213, + "grad_norm": 1.7104756832122803, + "learning_rate": 0.00018445435827041867, + "loss": 0.2024, + "step": 14950 + }, + { + "epoch": 0.23335621139327384, + "grad_norm": 0.4547693133354187, + "learning_rate": 0.0001844439591522639, + "loss": 0.271, + "step": 14960 + }, + { + "epoch": 0.23351219816559557, + "grad_norm": 4.291823863983154, + "learning_rate": 0.0001844335600341091, + "loss": 0.2195, + "step": 14970 + }, + { + "epoch": 0.23366818493791727, + "grad_norm": 2.0620875358581543, + "learning_rate": 0.00018442316091595433, + "loss": 0.2039, + "step": 14980 + }, + { + "epoch": 0.23382417171023898, + "grad_norm": 2.258686065673828, + "learning_rate": 0.00018441276179779955, + "loss": 0.3042, + "step": 14990 + }, + { + "epoch": 0.23398015848256068, + "grad_norm": 4.225653171539307, + "learning_rate": 0.00018440236267964477, + "loss": 0.2347, + "step": 15000 + }, + { + "epoch": 0.23413614525488238, + "grad_norm": 0.8725171685218811, + "learning_rate": 0.00018439196356149, + "loss": 0.3602, + "step": 15010 + }, + { + "epoch": 0.2342921320272041, + "grad_norm": 2.8600401878356934, + "learning_rate": 0.0001843815644433352, + "loss": 0.3224, + "step": 15020 + }, + { + "epoch": 0.2344481187995258, + "grad_norm": 2.8669276237487793, + "learning_rate": 0.00018437116532518043, + "loss": 0.2309, + "step": 15030 + }, + { + "epoch": 0.23460410557184752, + "grad_norm": 1.859756350517273, + "learning_rate": 0.00018436076620702565, + "loss": 0.3616, + "step": 15040 + }, + { + "epoch": 0.23476009234416922, + "grad_norm": 1.6354097127914429, + "learning_rate": 0.0001843503670888709, + "loss": 0.3151, + "step": 15050 + }, + { + "epoch": 0.23491607911649093, + "grad_norm": 1.152793288230896, + "learning_rate": 0.00018433996797071608, + "loss": 0.4332, + "step": 15060 + }, + { + "epoch": 0.23507206588881263, + "grad_norm": 2.024003744125366, + "learning_rate": 0.00018432956885256133, + "loss": 0.3026, + "step": 15070 + }, + { + "epoch": 0.23522805266113433, + "grad_norm": 0.8429051041603088, + "learning_rate": 0.00018431916973440652, + "loss": 0.1722, + "step": 15080 + }, + { + "epoch": 0.23538403943345604, + "grad_norm": 4.509194850921631, + "learning_rate": 0.00018430877061625177, + "loss": 0.2398, + "step": 15090 + }, + { + "epoch": 0.23554002620577774, + "grad_norm": 1.6877334117889404, + "learning_rate": 0.00018429837149809696, + "loss": 0.257, + "step": 15100 + }, + { + "epoch": 0.23569601297809945, + "grad_norm": 2.2232556343078613, + "learning_rate": 0.0001842879723799422, + "loss": 0.1704, + "step": 15110 + }, + { + "epoch": 0.23585199975042118, + "grad_norm": 1.5326635837554932, + "learning_rate": 0.0001842775732617874, + "loss": 0.2353, + "step": 15120 + }, + { + "epoch": 0.23600798652274288, + "grad_norm": 2.4672470092773438, + "learning_rate": 0.00018426717414363265, + "loss": 0.4524, + "step": 15130 + }, + { + "epoch": 0.23616397329506458, + "grad_norm": 1.8358993530273438, + "learning_rate": 0.00018425677502547784, + "loss": 0.3348, + "step": 15140 + }, + { + "epoch": 0.2363199600673863, + "grad_norm": 2.228985071182251, + "learning_rate": 0.0001842463759073231, + "loss": 0.5338, + "step": 15150 + }, + { + "epoch": 0.236475946839708, + "grad_norm": 2.7604992389678955, + "learning_rate": 0.00018423597678916828, + "loss": 0.2754, + "step": 15160 + }, + { + "epoch": 0.2366319336120297, + "grad_norm": 0.49068722128868103, + "learning_rate": 0.00018422557767101353, + "loss": 0.474, + "step": 15170 + }, + { + "epoch": 0.2367879203843514, + "grad_norm": 0.47539371252059937, + "learning_rate": 0.00018421517855285872, + "loss": 0.1101, + "step": 15180 + }, + { + "epoch": 0.2369439071566731, + "grad_norm": 0.5464065074920654, + "learning_rate": 0.00018420477943470397, + "loss": 0.1999, + "step": 15190 + }, + { + "epoch": 0.23709989392899483, + "grad_norm": 1.3112478256225586, + "learning_rate": 0.00018419438031654916, + "loss": 0.1641, + "step": 15200 + }, + { + "epoch": 0.23725588070131653, + "grad_norm": 0.14866875112056732, + "learning_rate": 0.0001841839811983944, + "loss": 0.2235, + "step": 15210 + }, + { + "epoch": 0.23741186747363824, + "grad_norm": 1.1525793075561523, + "learning_rate": 0.0001841735820802396, + "loss": 0.2579, + "step": 15220 + }, + { + "epoch": 0.23756785424595994, + "grad_norm": 2.644186019897461, + "learning_rate": 0.00018416318296208485, + "loss": 0.2991, + "step": 15230 + }, + { + "epoch": 0.23772384101828165, + "grad_norm": 1.0225379467010498, + "learning_rate": 0.00018415278384393004, + "loss": 0.1636, + "step": 15240 + }, + { + "epoch": 0.23787982779060335, + "grad_norm": 1.1804742813110352, + "learning_rate": 0.00018414238472577526, + "loss": 0.2387, + "step": 15250 + }, + { + "epoch": 0.23803581456292505, + "grad_norm": 2.0023930072784424, + "learning_rate": 0.00018413198560762048, + "loss": 0.1747, + "step": 15260 + }, + { + "epoch": 0.23819180133524678, + "grad_norm": 3.9979522228240967, + "learning_rate": 0.0001841215864894657, + "loss": 0.4135, + "step": 15270 + }, + { + "epoch": 0.2383477881075685, + "grad_norm": 2.5441789627075195, + "learning_rate": 0.00018411118737131092, + "loss": 0.3631, + "step": 15280 + }, + { + "epoch": 0.2385037748798902, + "grad_norm": 3.1089634895324707, + "learning_rate": 0.00018410078825315614, + "loss": 0.2644, + "step": 15290 + }, + { + "epoch": 0.2386597616522119, + "grad_norm": 0.0912027433514595, + "learning_rate": 0.00018409038913500136, + "loss": 0.1877, + "step": 15300 + }, + { + "epoch": 0.2388157484245336, + "grad_norm": 2.1551854610443115, + "learning_rate": 0.00018407999001684658, + "loss": 0.1639, + "step": 15310 + }, + { + "epoch": 0.2389717351968553, + "grad_norm": 0.25086110830307007, + "learning_rate": 0.0001840695908986918, + "loss": 0.2516, + "step": 15320 + }, + { + "epoch": 0.239127721969177, + "grad_norm": 3.022775888442993, + "learning_rate": 0.00018405919178053702, + "loss": 0.5105, + "step": 15330 + }, + { + "epoch": 0.2392837087414987, + "grad_norm": 1.9660505056381226, + "learning_rate": 0.00018404879266238223, + "loss": 0.2749, + "step": 15340 + }, + { + "epoch": 0.23943969551382044, + "grad_norm": 1.7173848152160645, + "learning_rate": 0.00018403839354422745, + "loss": 0.3761, + "step": 15350 + }, + { + "epoch": 0.23959568228614214, + "grad_norm": 2.349210023880005, + "learning_rate": 0.00018402799442607267, + "loss": 0.3338, + "step": 15360 + }, + { + "epoch": 0.23975166905846385, + "grad_norm": 0.2212291657924652, + "learning_rate": 0.0001840175953079179, + "loss": 0.4127, + "step": 15370 + }, + { + "epoch": 0.23990765583078555, + "grad_norm": 0.2859221398830414, + "learning_rate": 0.0001840071961897631, + "loss": 0.3477, + "step": 15380 + }, + { + "epoch": 0.24006364260310725, + "grad_norm": 0.4488182067871094, + "learning_rate": 0.00018399679707160833, + "loss": 0.1695, + "step": 15390 + }, + { + "epoch": 0.24021962937542896, + "grad_norm": 1.323156714439392, + "learning_rate": 0.00018398639795345355, + "loss": 0.3143, + "step": 15400 + }, + { + "epoch": 0.24037561614775066, + "grad_norm": 0.8333446979522705, + "learning_rate": 0.00018397599883529877, + "loss": 0.3662, + "step": 15410 + }, + { + "epoch": 0.2405316029200724, + "grad_norm": 1.69521164894104, + "learning_rate": 0.000183965599717144, + "loss": 0.1598, + "step": 15420 + }, + { + "epoch": 0.2406875896923941, + "grad_norm": 2.484534740447998, + "learning_rate": 0.0001839552005989892, + "loss": 0.2692, + "step": 15430 + }, + { + "epoch": 0.2408435764647158, + "grad_norm": 1.367640733718872, + "learning_rate": 0.00018394480148083443, + "loss": 0.4109, + "step": 15440 + }, + { + "epoch": 0.2409995632370375, + "grad_norm": 0.6717288494110107, + "learning_rate": 0.00018393440236267965, + "loss": 0.2132, + "step": 15450 + }, + { + "epoch": 0.2411555500093592, + "grad_norm": 0.7880472540855408, + "learning_rate": 0.00018392400324452487, + "loss": 0.2627, + "step": 15460 + }, + { + "epoch": 0.2413115367816809, + "grad_norm": 0.807041347026825, + "learning_rate": 0.0001839136041263701, + "loss": 0.3131, + "step": 15470 + }, + { + "epoch": 0.2414675235540026, + "grad_norm": 4.352522850036621, + "learning_rate": 0.0001839032050082153, + "loss": 0.4084, + "step": 15480 + }, + { + "epoch": 0.24162351032632431, + "grad_norm": 4.362399101257324, + "learning_rate": 0.00018389280589006053, + "loss": 0.2234, + "step": 15490 + }, + { + "epoch": 0.24177949709864605, + "grad_norm": 5.924840450286865, + "learning_rate": 0.00018388240677190575, + "loss": 0.4518, + "step": 15500 + }, + { + "epoch": 0.24193548387096775, + "grad_norm": 0.28530246019363403, + "learning_rate": 0.00018387200765375097, + "loss": 0.2088, + "step": 15510 + }, + { + "epoch": 0.24209147064328945, + "grad_norm": 0.1666223257780075, + "learning_rate": 0.0001838616085355962, + "loss": 0.3483, + "step": 15520 + }, + { + "epoch": 0.24224745741561116, + "grad_norm": 4.683041095733643, + "learning_rate": 0.0001838512094174414, + "loss": 0.3149, + "step": 15530 + }, + { + "epoch": 0.24240344418793286, + "grad_norm": 0.6108630895614624, + "learning_rate": 0.00018384081029928663, + "loss": 0.3006, + "step": 15540 + }, + { + "epoch": 0.24255943096025456, + "grad_norm": 0.9650145173072815, + "learning_rate": 0.00018383041118113185, + "loss": 0.1056, + "step": 15550 + }, + { + "epoch": 0.24271541773257627, + "grad_norm": 1.0948867797851562, + "learning_rate": 0.00018382001206297707, + "loss": 0.2872, + "step": 15560 + }, + { + "epoch": 0.242871404504898, + "grad_norm": 1.5407096147537231, + "learning_rate": 0.0001838096129448223, + "loss": 0.328, + "step": 15570 + }, + { + "epoch": 0.2430273912772197, + "grad_norm": 5.225300312042236, + "learning_rate": 0.0001837992138266675, + "loss": 0.2459, + "step": 15580 + }, + { + "epoch": 0.2431833780495414, + "grad_norm": 1.5624916553497314, + "learning_rate": 0.00018378881470851273, + "loss": 0.317, + "step": 15590 + }, + { + "epoch": 0.2433393648218631, + "grad_norm": 2.365774154663086, + "learning_rate": 0.00018377841559035795, + "loss": 0.2278, + "step": 15600 + }, + { + "epoch": 0.2434953515941848, + "grad_norm": 2.1035828590393066, + "learning_rate": 0.00018376801647220317, + "loss": 0.4028, + "step": 15610 + }, + { + "epoch": 0.24365133836650651, + "grad_norm": 0.13403718173503876, + "learning_rate": 0.00018375761735404838, + "loss": 0.2898, + "step": 15620 + }, + { + "epoch": 0.24380732513882822, + "grad_norm": 0.6040320992469788, + "learning_rate": 0.0001837472182358936, + "loss": 0.316, + "step": 15630 + }, + { + "epoch": 0.24396331191114992, + "grad_norm": 1.187819242477417, + "learning_rate": 0.00018373681911773882, + "loss": 0.3436, + "step": 15640 + }, + { + "epoch": 0.24411929868347165, + "grad_norm": 0.8004405498504639, + "learning_rate": 0.00018372641999958404, + "loss": 0.1658, + "step": 15650 + }, + { + "epoch": 0.24427528545579336, + "grad_norm": 1.1902847290039062, + "learning_rate": 0.00018371602088142926, + "loss": 0.3159, + "step": 15660 + }, + { + "epoch": 0.24443127222811506, + "grad_norm": 2.7241697311401367, + "learning_rate": 0.00018370562176327448, + "loss": 0.3026, + "step": 15670 + }, + { + "epoch": 0.24458725900043676, + "grad_norm": 1.1655789613723755, + "learning_rate": 0.0001836952226451197, + "loss": 0.1547, + "step": 15680 + }, + { + "epoch": 0.24474324577275847, + "grad_norm": 2.054489850997925, + "learning_rate": 0.00018368482352696492, + "loss": 0.2555, + "step": 15690 + }, + { + "epoch": 0.24489923254508017, + "grad_norm": 0.14498686790466309, + "learning_rate": 0.00018367442440881014, + "loss": 0.1483, + "step": 15700 + }, + { + "epoch": 0.24505521931740187, + "grad_norm": 1.7551847696304321, + "learning_rate": 0.00018366402529065536, + "loss": 0.4449, + "step": 15710 + }, + { + "epoch": 0.2452112060897236, + "grad_norm": 2.135995388031006, + "learning_rate": 0.00018365362617250058, + "loss": 0.3155, + "step": 15720 + }, + { + "epoch": 0.2453671928620453, + "grad_norm": 1.8099571466445923, + "learning_rate": 0.0001836432270543458, + "loss": 0.4421, + "step": 15730 + }, + { + "epoch": 0.245523179634367, + "grad_norm": 1.265748381614685, + "learning_rate": 0.00018363282793619102, + "loss": 0.197, + "step": 15740 + }, + { + "epoch": 0.24567916640668871, + "grad_norm": 0.577680230140686, + "learning_rate": 0.00018362242881803624, + "loss": 0.1765, + "step": 15750 + }, + { + "epoch": 0.24583515317901042, + "grad_norm": 1.6049988269805908, + "learning_rate": 0.00018361202969988146, + "loss": 0.2905, + "step": 15760 + }, + { + "epoch": 0.24599113995133212, + "grad_norm": 1.461227297782898, + "learning_rate": 0.00018360163058172668, + "loss": 0.3608, + "step": 15770 + }, + { + "epoch": 0.24614712672365383, + "grad_norm": 2.4718856811523438, + "learning_rate": 0.0001835912314635719, + "loss": 0.3934, + "step": 15780 + }, + { + "epoch": 0.24630311349597553, + "grad_norm": 2.033041477203369, + "learning_rate": 0.00018358083234541712, + "loss": 0.1819, + "step": 15790 + }, + { + "epoch": 0.24645910026829726, + "grad_norm": 0.011789345182478428, + "learning_rate": 0.00018357043322726234, + "loss": 0.3887, + "step": 15800 + }, + { + "epoch": 0.24661508704061896, + "grad_norm": 1.5813320875167847, + "learning_rate": 0.00018356003410910756, + "loss": 0.2348, + "step": 15810 + }, + { + "epoch": 0.24677107381294067, + "grad_norm": 6.326842784881592, + "learning_rate": 0.00018354963499095278, + "loss": 0.2567, + "step": 15820 + }, + { + "epoch": 0.24692706058526237, + "grad_norm": 3.0133228302001953, + "learning_rate": 0.000183539235872798, + "loss": 0.4596, + "step": 15830 + }, + { + "epoch": 0.24708304735758407, + "grad_norm": 0.8021423816680908, + "learning_rate": 0.00018352883675464322, + "loss": 0.4769, + "step": 15840 + }, + { + "epoch": 0.24723903412990578, + "grad_norm": 1.221909999847412, + "learning_rate": 0.00018351843763648844, + "loss": 0.1225, + "step": 15850 + }, + { + "epoch": 0.24739502090222748, + "grad_norm": 1.2123382091522217, + "learning_rate": 0.00018350803851833366, + "loss": 0.4187, + "step": 15860 + }, + { + "epoch": 0.2475510076745492, + "grad_norm": 0.9829211235046387, + "learning_rate": 0.00018349763940017888, + "loss": 0.2008, + "step": 15870 + }, + { + "epoch": 0.24770699444687092, + "grad_norm": 0.16013558208942413, + "learning_rate": 0.0001834872402820241, + "loss": 0.3002, + "step": 15880 + }, + { + "epoch": 0.24786298121919262, + "grad_norm": 2.275714874267578, + "learning_rate": 0.00018347684116386932, + "loss": 0.3606, + "step": 15890 + }, + { + "epoch": 0.24801896799151432, + "grad_norm": 0.4143832325935364, + "learning_rate": 0.00018346644204571453, + "loss": 0.3834, + "step": 15900 + }, + { + "epoch": 0.24817495476383603, + "grad_norm": 2.2170028686523438, + "learning_rate": 0.00018345604292755975, + "loss": 0.2724, + "step": 15910 + }, + { + "epoch": 0.24833094153615773, + "grad_norm": 0.5369767546653748, + "learning_rate": 0.00018344564380940497, + "loss": 0.3011, + "step": 15920 + }, + { + "epoch": 0.24848692830847943, + "grad_norm": 2.0653202533721924, + "learning_rate": 0.0001834352446912502, + "loss": 0.3613, + "step": 15930 + }, + { + "epoch": 0.24864291508080114, + "grad_norm": 1.3467276096343994, + "learning_rate": 0.0001834248455730954, + "loss": 0.1488, + "step": 15940 + }, + { + "epoch": 0.24879890185312287, + "grad_norm": 1.5526561737060547, + "learning_rate": 0.00018341444645494063, + "loss": 0.2715, + "step": 15950 + }, + { + "epoch": 0.24895488862544457, + "grad_norm": 0.7290893793106079, + "learning_rate": 0.00018340404733678585, + "loss": 0.2292, + "step": 15960 + }, + { + "epoch": 0.24911087539776627, + "grad_norm": 1.8766149282455444, + "learning_rate": 0.00018339364821863107, + "loss": 0.3524, + "step": 15970 + }, + { + "epoch": 0.24926686217008798, + "grad_norm": 0.400741308927536, + "learning_rate": 0.0001833832491004763, + "loss": 0.2937, + "step": 15980 + }, + { + "epoch": 0.24942284894240968, + "grad_norm": 0.8890944719314575, + "learning_rate": 0.0001833728499823215, + "loss": 0.2384, + "step": 15990 + }, + { + "epoch": 0.24957883571473138, + "grad_norm": 2.274386405944824, + "learning_rate": 0.00018336245086416673, + "loss": 0.2967, + "step": 16000 + }, + { + "epoch": 0.2497348224870531, + "grad_norm": 0.5878872275352478, + "learning_rate": 0.00018335205174601195, + "loss": 0.44, + "step": 16010 + }, + { + "epoch": 0.24989080925937482, + "grad_norm": 2.2065277099609375, + "learning_rate": 0.00018334165262785717, + "loss": 0.3069, + "step": 16020 + }, + { + "epoch": 0.2500467960316965, + "grad_norm": 2.53945255279541, + "learning_rate": 0.0001833312535097024, + "loss": 0.2524, + "step": 16030 + }, + { + "epoch": 0.2502027828040182, + "grad_norm": 0.961932361125946, + "learning_rate": 0.0001833208543915476, + "loss": 0.1822, + "step": 16040 + }, + { + "epoch": 0.2503587695763399, + "grad_norm": 2.5592494010925293, + "learning_rate": 0.00018331045527339283, + "loss": 0.1402, + "step": 16050 + }, + { + "epoch": 0.25051475634866166, + "grad_norm": 0.5766373872756958, + "learning_rate": 0.00018330005615523805, + "loss": 0.4622, + "step": 16060 + }, + { + "epoch": 0.25067074312098336, + "grad_norm": 0.7966228723526001, + "learning_rate": 0.00018328965703708327, + "loss": 0.5342, + "step": 16070 + }, + { + "epoch": 0.25082672989330507, + "grad_norm": 2.3399205207824707, + "learning_rate": 0.0001832792579189285, + "loss": 0.3567, + "step": 16080 + }, + { + "epoch": 0.25098271666562677, + "grad_norm": 1.6496340036392212, + "learning_rate": 0.0001832688588007737, + "loss": 0.4204, + "step": 16090 + }, + { + "epoch": 0.2511387034379485, + "grad_norm": 0.5356587171554565, + "learning_rate": 0.00018325845968261893, + "loss": 0.2006, + "step": 16100 + }, + { + "epoch": 0.2512946902102702, + "grad_norm": 1.7309391498565674, + "learning_rate": 0.00018324806056446415, + "loss": 0.328, + "step": 16110 + }, + { + "epoch": 0.2514506769825919, + "grad_norm": 0.15209828317165375, + "learning_rate": 0.00018323766144630937, + "loss": 0.2388, + "step": 16120 + }, + { + "epoch": 0.2516066637549136, + "grad_norm": 0.1848757416009903, + "learning_rate": 0.0001832272623281546, + "loss": 0.2857, + "step": 16130 + }, + { + "epoch": 0.2517626505272353, + "grad_norm": 2.939927816390991, + "learning_rate": 0.0001832168632099998, + "loss": 0.1881, + "step": 16140 + }, + { + "epoch": 0.251918637299557, + "grad_norm": 7.412752628326416, + "learning_rate": 0.000183206464091845, + "loss": 0.4281, + "step": 16150 + }, + { + "epoch": 0.2520746240718787, + "grad_norm": 2.9228932857513428, + "learning_rate": 0.00018319606497369025, + "loss": 0.4912, + "step": 16160 + }, + { + "epoch": 0.2522306108442004, + "grad_norm": 0.21853139996528625, + "learning_rate": 0.00018318566585553544, + "loss": 0.2368, + "step": 16170 + }, + { + "epoch": 0.2523865976165221, + "grad_norm": 2.6338698863983154, + "learning_rate": 0.00018317526673738068, + "loss": 0.1808, + "step": 16180 + }, + { + "epoch": 0.2525425843888438, + "grad_norm": 0.6143313646316528, + "learning_rate": 0.00018316486761922588, + "loss": 0.222, + "step": 16190 + }, + { + "epoch": 0.2526985711611655, + "grad_norm": 2.7535603046417236, + "learning_rate": 0.00018315446850107112, + "loss": 0.3118, + "step": 16200 + }, + { + "epoch": 0.25285455793348727, + "grad_norm": 1.3380391597747803, + "learning_rate": 0.00018314406938291632, + "loss": 0.3336, + "step": 16210 + }, + { + "epoch": 0.25301054470580897, + "grad_norm": 0.8309090733528137, + "learning_rate": 0.00018313367026476156, + "loss": 0.2797, + "step": 16220 + }, + { + "epoch": 0.2531665314781307, + "grad_norm": 2.4775400161743164, + "learning_rate": 0.00018312327114660676, + "loss": 0.397, + "step": 16230 + }, + { + "epoch": 0.2533225182504524, + "grad_norm": 0.2340003252029419, + "learning_rate": 0.000183112872028452, + "loss": 0.1818, + "step": 16240 + }, + { + "epoch": 0.2534785050227741, + "grad_norm": 0.26473724842071533, + "learning_rate": 0.0001831024729102972, + "loss": 0.1421, + "step": 16250 + }, + { + "epoch": 0.2536344917950958, + "grad_norm": 2.526986837387085, + "learning_rate": 0.00018309207379214244, + "loss": 0.2368, + "step": 16260 + }, + { + "epoch": 0.2537904785674175, + "grad_norm": 1.5454362630844116, + "learning_rate": 0.00018308167467398763, + "loss": 0.1777, + "step": 16270 + }, + { + "epoch": 0.2539464653397392, + "grad_norm": 1.227616548538208, + "learning_rate": 0.00018307127555583288, + "loss": 0.3608, + "step": 16280 + }, + { + "epoch": 0.2541024521120609, + "grad_norm": 0.6660637259483337, + "learning_rate": 0.00018306087643767807, + "loss": 0.3502, + "step": 16290 + }, + { + "epoch": 0.2542584388843826, + "grad_norm": 2.6408071517944336, + "learning_rate": 0.00018305047731952332, + "loss": 0.2044, + "step": 16300 + }, + { + "epoch": 0.2544144256567043, + "grad_norm": 3.001298427581787, + "learning_rate": 0.0001830400782013685, + "loss": 0.261, + "step": 16310 + }, + { + "epoch": 0.254570412429026, + "grad_norm": 2.0070767402648926, + "learning_rate": 0.00018302967908321376, + "loss": 0.74, + "step": 16320 + }, + { + "epoch": 0.2547263992013477, + "grad_norm": 0.3652030825614929, + "learning_rate": 0.00018301927996505895, + "loss": 0.1292, + "step": 16330 + }, + { + "epoch": 0.2548823859736694, + "grad_norm": 0.7860630750656128, + "learning_rate": 0.0001830088808469042, + "loss": 0.1655, + "step": 16340 + }, + { + "epoch": 0.2550383727459911, + "grad_norm": 1.5533596277236938, + "learning_rate": 0.0001829984817287494, + "loss": 0.1662, + "step": 16350 + }, + { + "epoch": 0.2551943595183129, + "grad_norm": 0.5608229041099548, + "learning_rate": 0.00018298808261059464, + "loss": 0.2899, + "step": 16360 + }, + { + "epoch": 0.2553503462906346, + "grad_norm": 1.461485505104065, + "learning_rate": 0.00018297768349243983, + "loss": 0.2443, + "step": 16370 + }, + { + "epoch": 0.2555063330629563, + "grad_norm": 3.218679666519165, + "learning_rate": 0.00018296728437428508, + "loss": 0.2234, + "step": 16380 + }, + { + "epoch": 0.255662319835278, + "grad_norm": 0.6378768682479858, + "learning_rate": 0.0001829568852561303, + "loss": 0.23, + "step": 16390 + }, + { + "epoch": 0.2558183066075997, + "grad_norm": 1.5998066663742065, + "learning_rate": 0.00018294648613797552, + "loss": 0.3581, + "step": 16400 + }, + { + "epoch": 0.2559742933799214, + "grad_norm": 3.643050193786621, + "learning_rate": 0.00018293608701982074, + "loss": 0.4532, + "step": 16410 + }, + { + "epoch": 0.2561302801522431, + "grad_norm": 2.9357807636260986, + "learning_rate": 0.00018292568790166596, + "loss": 0.3143, + "step": 16420 + }, + { + "epoch": 0.2562862669245648, + "grad_norm": 4.367801666259766, + "learning_rate": 0.00018291528878351118, + "loss": 0.3181, + "step": 16430 + }, + { + "epoch": 0.2564422536968865, + "grad_norm": 1.823776125907898, + "learning_rate": 0.0001829048896653564, + "loss": 0.3028, + "step": 16440 + }, + { + "epoch": 0.2565982404692082, + "grad_norm": 1.0017749071121216, + "learning_rate": 0.00018289449054720162, + "loss": 0.282, + "step": 16450 + }, + { + "epoch": 0.2567542272415299, + "grad_norm": 3.6332972049713135, + "learning_rate": 0.00018288409142904683, + "loss": 0.1886, + "step": 16460 + }, + { + "epoch": 0.2569102140138516, + "grad_norm": 2.792665719985962, + "learning_rate": 0.00018287369231089205, + "loss": 0.2458, + "step": 16470 + }, + { + "epoch": 0.2570662007861733, + "grad_norm": 1.8526560068130493, + "learning_rate": 0.00018286329319273727, + "loss": 0.3822, + "step": 16480 + }, + { + "epoch": 0.257222187558495, + "grad_norm": 1.481366515159607, + "learning_rate": 0.0001828528940745825, + "loss": 0.4398, + "step": 16490 + }, + { + "epoch": 0.2573781743308167, + "grad_norm": 1.8167740106582642, + "learning_rate": 0.0001828424949564277, + "loss": 0.2492, + "step": 16500 + }, + { + "epoch": 0.2575341611031385, + "grad_norm": 4.7140936851501465, + "learning_rate": 0.00018283209583827293, + "loss": 0.3269, + "step": 16510 + }, + { + "epoch": 0.2576901478754602, + "grad_norm": 4.104098796844482, + "learning_rate": 0.00018282169672011815, + "loss": 0.2107, + "step": 16520 + }, + { + "epoch": 0.2578461346477819, + "grad_norm": 2.7594456672668457, + "learning_rate": 0.00018281129760196337, + "loss": 0.2852, + "step": 16530 + }, + { + "epoch": 0.2580021214201036, + "grad_norm": 2.0960562229156494, + "learning_rate": 0.0001828008984838086, + "loss": 0.276, + "step": 16540 + }, + { + "epoch": 0.2581581081924253, + "grad_norm": 1.2023634910583496, + "learning_rate": 0.0001827904993656538, + "loss": 0.1824, + "step": 16550 + }, + { + "epoch": 0.258314094964747, + "grad_norm": 1.159704566001892, + "learning_rate": 0.00018278010024749903, + "loss": 0.2108, + "step": 16560 + }, + { + "epoch": 0.2584700817370687, + "grad_norm": 0.6380292773246765, + "learning_rate": 0.00018276970112934425, + "loss": 0.2586, + "step": 16570 + }, + { + "epoch": 0.2586260685093904, + "grad_norm": 0.05646060034632683, + "learning_rate": 0.00018275930201118947, + "loss": 0.4047, + "step": 16580 + }, + { + "epoch": 0.2587820552817121, + "grad_norm": 3.3459885120391846, + "learning_rate": 0.0001827489028930347, + "loss": 0.2384, + "step": 16590 + }, + { + "epoch": 0.2589380420540338, + "grad_norm": 1.6905351877212524, + "learning_rate": 0.0001827385037748799, + "loss": 0.3039, + "step": 16600 + }, + { + "epoch": 0.2590940288263555, + "grad_norm": 1.4395368099212646, + "learning_rate": 0.00018272810465672513, + "loss": 0.2976, + "step": 16610 + }, + { + "epoch": 0.2592500155986772, + "grad_norm": 7.715827465057373, + "learning_rate": 0.00018271770553857032, + "loss": 0.2334, + "step": 16620 + }, + { + "epoch": 0.2594060023709989, + "grad_norm": 4.04732608795166, + "learning_rate": 0.00018270730642041557, + "loss": 0.2251, + "step": 16630 + }, + { + "epoch": 0.2595619891433206, + "grad_norm": 0.1727902591228485, + "learning_rate": 0.00018269690730226076, + "loss": 0.1998, + "step": 16640 + }, + { + "epoch": 0.25971797591564233, + "grad_norm": 2.7462995052337646, + "learning_rate": 0.000182686508184106, + "loss": 0.2641, + "step": 16650 + }, + { + "epoch": 0.25987396268796403, + "grad_norm": 3.3414735794067383, + "learning_rate": 0.0001826761090659512, + "loss": 0.2143, + "step": 16660 + }, + { + "epoch": 0.2600299494602858, + "grad_norm": 2.718449354171753, + "learning_rate": 0.00018266570994779645, + "loss": 0.1722, + "step": 16670 + }, + { + "epoch": 0.2601859362326075, + "grad_norm": 0.7628939151763916, + "learning_rate": 0.00018265531082964164, + "loss": 0.2568, + "step": 16680 + }, + { + "epoch": 0.2603419230049292, + "grad_norm": 3.6897547245025635, + "learning_rate": 0.00018264491171148689, + "loss": 0.2673, + "step": 16690 + }, + { + "epoch": 0.2604979097772509, + "grad_norm": 6.371345520019531, + "learning_rate": 0.00018263451259333208, + "loss": 0.263, + "step": 16700 + }, + { + "epoch": 0.2606538965495726, + "grad_norm": 1.7923188209533691, + "learning_rate": 0.00018262411347517733, + "loss": 0.3099, + "step": 16710 + }, + { + "epoch": 0.2608098833218943, + "grad_norm": 0.5189704895019531, + "learning_rate": 0.00018261371435702252, + "loss": 0.1695, + "step": 16720 + }, + { + "epoch": 0.260965870094216, + "grad_norm": 3.3882267475128174, + "learning_rate": 0.00018260331523886776, + "loss": 0.3388, + "step": 16730 + }, + { + "epoch": 0.2611218568665377, + "grad_norm": 0.8821435570716858, + "learning_rate": 0.00018259291612071296, + "loss": 0.1749, + "step": 16740 + }, + { + "epoch": 0.2612778436388594, + "grad_norm": 4.573073387145996, + "learning_rate": 0.0001825825170025582, + "loss": 0.353, + "step": 16750 + }, + { + "epoch": 0.2614338304111811, + "grad_norm": 0.11496096104383469, + "learning_rate": 0.0001825721178844034, + "loss": 0.4176, + "step": 16760 + }, + { + "epoch": 0.2615898171835028, + "grad_norm": 0.6777880191802979, + "learning_rate": 0.00018256171876624864, + "loss": 0.1907, + "step": 16770 + }, + { + "epoch": 0.26174580395582453, + "grad_norm": 2.2113587856292725, + "learning_rate": 0.00018255131964809384, + "loss": 0.2148, + "step": 16780 + }, + { + "epoch": 0.26190179072814623, + "grad_norm": 0.8497506380081177, + "learning_rate": 0.00018254092052993908, + "loss": 0.5211, + "step": 16790 + }, + { + "epoch": 0.26205777750046794, + "grad_norm": 2.6945035457611084, + "learning_rate": 0.00018253052141178428, + "loss": 0.3768, + "step": 16800 + }, + { + "epoch": 0.26221376427278964, + "grad_norm": 0.5304585099220276, + "learning_rate": 0.00018252012229362952, + "loss": 0.194, + "step": 16810 + }, + { + "epoch": 0.2623697510451114, + "grad_norm": 0.25179457664489746, + "learning_rate": 0.00018250972317547471, + "loss": 0.4074, + "step": 16820 + }, + { + "epoch": 0.2625257378174331, + "grad_norm": 2.897371768951416, + "learning_rate": 0.00018249932405731996, + "loss": 0.2992, + "step": 16830 + }, + { + "epoch": 0.2626817245897548, + "grad_norm": 1.3298170566558838, + "learning_rate": 0.00018248892493916515, + "loss": 0.2253, + "step": 16840 + }, + { + "epoch": 0.2628377113620765, + "grad_norm": 1.2430392503738403, + "learning_rate": 0.0001824785258210104, + "loss": 0.2164, + "step": 16850 + }, + { + "epoch": 0.2629936981343982, + "grad_norm": 1.117514967918396, + "learning_rate": 0.0001824681267028556, + "loss": 0.2139, + "step": 16860 + }, + { + "epoch": 0.2631496849067199, + "grad_norm": 0.6369630694389343, + "learning_rate": 0.00018245772758470084, + "loss": 0.2957, + "step": 16870 + }, + { + "epoch": 0.2633056716790416, + "grad_norm": 3.72832989692688, + "learning_rate": 0.00018244732846654603, + "loss": 0.2913, + "step": 16880 + }, + { + "epoch": 0.2634616584513633, + "grad_norm": 1.336425542831421, + "learning_rate": 0.00018243692934839128, + "loss": 0.2601, + "step": 16890 + }, + { + "epoch": 0.263617645223685, + "grad_norm": 0.29804128408432007, + "learning_rate": 0.00018242653023023647, + "loss": 0.1844, + "step": 16900 + }, + { + "epoch": 0.26377363199600673, + "grad_norm": 0.7500967383384705, + "learning_rate": 0.00018241613111208172, + "loss": 0.3826, + "step": 16910 + }, + { + "epoch": 0.26392961876832843, + "grad_norm": 1.267012119293213, + "learning_rate": 0.0001824057319939269, + "loss": 0.3804, + "step": 16920 + }, + { + "epoch": 0.26408560554065014, + "grad_norm": 0.9892479181289673, + "learning_rate": 0.00018239533287577216, + "loss": 0.3588, + "step": 16930 + }, + { + "epoch": 0.26424159231297184, + "grad_norm": 3.285210132598877, + "learning_rate": 0.00018238493375761735, + "loss": 0.2827, + "step": 16940 + }, + { + "epoch": 0.26439757908529354, + "grad_norm": 0.42420023679733276, + "learning_rate": 0.0001823745346394626, + "loss": 0.4238, + "step": 16950 + }, + { + "epoch": 0.26455356585761525, + "grad_norm": 2.081674098968506, + "learning_rate": 0.0001823641355213078, + "loss": 0.2595, + "step": 16960 + }, + { + "epoch": 0.264709552629937, + "grad_norm": 3.789632558822632, + "learning_rate": 0.00018235373640315304, + "loss": 0.2646, + "step": 16970 + }, + { + "epoch": 0.2648655394022587, + "grad_norm": 8.834197044372559, + "learning_rate": 0.00018234333728499823, + "loss": 0.3295, + "step": 16980 + }, + { + "epoch": 0.2650215261745804, + "grad_norm": 3.3150501251220703, + "learning_rate": 0.00018233293816684348, + "loss": 0.3891, + "step": 16990 + }, + { + "epoch": 0.2651775129469021, + "grad_norm": 1.3610211610794067, + "learning_rate": 0.00018232253904868867, + "loss": 0.3027, + "step": 17000 + }, + { + "epoch": 0.2653334997192238, + "grad_norm": 1.1531686782836914, + "learning_rate": 0.00018231213993053391, + "loss": 0.1004, + "step": 17010 + }, + { + "epoch": 0.2654894864915455, + "grad_norm": 0.9656435251235962, + "learning_rate": 0.0001823017408123791, + "loss": 0.1935, + "step": 17020 + }, + { + "epoch": 0.2656454732638672, + "grad_norm": 1.1924324035644531, + "learning_rate": 0.00018229134169422435, + "loss": 0.2528, + "step": 17030 + }, + { + "epoch": 0.26580146003618893, + "grad_norm": 6.551888942718506, + "learning_rate": 0.00018228094257606955, + "loss": 0.3801, + "step": 17040 + }, + { + "epoch": 0.26595744680851063, + "grad_norm": 3.5191943645477295, + "learning_rate": 0.0001822705434579148, + "loss": 0.4463, + "step": 17050 + }, + { + "epoch": 0.26611343358083234, + "grad_norm": 0.6895869970321655, + "learning_rate": 0.00018226014433975999, + "loss": 0.1161, + "step": 17060 + }, + { + "epoch": 0.26626942035315404, + "grad_norm": 2.485004186630249, + "learning_rate": 0.0001822497452216052, + "loss": 0.3215, + "step": 17070 + }, + { + "epoch": 0.26642540712547574, + "grad_norm": 3.704139232635498, + "learning_rate": 0.00018223934610345043, + "loss": 0.1767, + "step": 17080 + }, + { + "epoch": 0.26658139389779745, + "grad_norm": 3.2014617919921875, + "learning_rate": 0.00018222894698529564, + "loss": 0.1705, + "step": 17090 + }, + { + "epoch": 0.26673738067011915, + "grad_norm": 0.45712214708328247, + "learning_rate": 0.00018221854786714086, + "loss": 0.2746, + "step": 17100 + }, + { + "epoch": 0.26689336744244085, + "grad_norm": 4.020730495452881, + "learning_rate": 0.00018220814874898608, + "loss": 0.2358, + "step": 17110 + }, + { + "epoch": 0.2670493542147626, + "grad_norm": 2.0267622470855713, + "learning_rate": 0.0001821977496308313, + "loss": 0.1414, + "step": 17120 + }, + { + "epoch": 0.2672053409870843, + "grad_norm": 1.721405029296875, + "learning_rate": 0.00018218735051267652, + "loss": 0.1746, + "step": 17130 + }, + { + "epoch": 0.267361327759406, + "grad_norm": 0.011413615196943283, + "learning_rate": 0.00018217695139452174, + "loss": 0.2259, + "step": 17140 + }, + { + "epoch": 0.2675173145317277, + "grad_norm": 0.08803238719701767, + "learning_rate": 0.00018216655227636696, + "loss": 0.2077, + "step": 17150 + }, + { + "epoch": 0.2676733013040494, + "grad_norm": 3.741621971130371, + "learning_rate": 0.00018215615315821218, + "loss": 0.3209, + "step": 17160 + }, + { + "epoch": 0.26782928807637113, + "grad_norm": 1.7220309972763062, + "learning_rate": 0.0001821457540400574, + "loss": 0.4369, + "step": 17170 + }, + { + "epoch": 0.26798527484869283, + "grad_norm": 2.22023344039917, + "learning_rate": 0.00018213535492190262, + "loss": 0.4086, + "step": 17180 + }, + { + "epoch": 0.26814126162101454, + "grad_norm": 0.4156394898891449, + "learning_rate": 0.00018212495580374784, + "loss": 0.0675, + "step": 17190 + }, + { + "epoch": 0.26829724839333624, + "grad_norm": 2.965853691101074, + "learning_rate": 0.00018211455668559306, + "loss": 0.3197, + "step": 17200 + }, + { + "epoch": 0.26845323516565794, + "grad_norm": 0.5918697714805603, + "learning_rate": 0.00018210415756743828, + "loss": 0.4966, + "step": 17210 + }, + { + "epoch": 0.26860922193797965, + "grad_norm": 2.0074431896209717, + "learning_rate": 0.0001820937584492835, + "loss": 0.5657, + "step": 17220 + }, + { + "epoch": 0.26876520871030135, + "grad_norm": 1.7945674657821655, + "learning_rate": 0.00018208335933112872, + "loss": 0.1773, + "step": 17230 + }, + { + "epoch": 0.26892119548262305, + "grad_norm": 2.769350528717041, + "learning_rate": 0.00018207296021297397, + "loss": 0.1756, + "step": 17240 + }, + { + "epoch": 0.26907718225494476, + "grad_norm": 2.0327277183532715, + "learning_rate": 0.00018206256109481916, + "loss": 0.2468, + "step": 17250 + }, + { + "epoch": 0.26923316902726646, + "grad_norm": 1.1871803998947144, + "learning_rate": 0.0001820521619766644, + "loss": 0.2707, + "step": 17260 + }, + { + "epoch": 0.2693891557995882, + "grad_norm": 1.248678207397461, + "learning_rate": 0.0001820417628585096, + "loss": 0.2328, + "step": 17270 + }, + { + "epoch": 0.2695451425719099, + "grad_norm": 1.1052762269973755, + "learning_rate": 0.00018203136374035485, + "loss": 0.5168, + "step": 17280 + }, + { + "epoch": 0.2697011293442316, + "grad_norm": 3.7844345569610596, + "learning_rate": 0.00018202096462220004, + "loss": 0.3097, + "step": 17290 + }, + { + "epoch": 0.26985711611655333, + "grad_norm": 1.474967122077942, + "learning_rate": 0.00018201056550404528, + "loss": 0.2622, + "step": 17300 + }, + { + "epoch": 0.27001310288887503, + "grad_norm": 1.0871832370758057, + "learning_rate": 0.00018200016638589048, + "loss": 0.3639, + "step": 17310 + }, + { + "epoch": 0.27016908966119674, + "grad_norm": 3.5728237628936768, + "learning_rate": 0.00018198976726773572, + "loss": 0.2158, + "step": 17320 + }, + { + "epoch": 0.27032507643351844, + "grad_norm": 2.015841245651245, + "learning_rate": 0.00018197936814958092, + "loss": 0.2148, + "step": 17330 + }, + { + "epoch": 0.27048106320584014, + "grad_norm": 3.2053275108337402, + "learning_rate": 0.00018196896903142616, + "loss": 0.4, + "step": 17340 + }, + { + "epoch": 0.27063704997816185, + "grad_norm": 3.5583112239837646, + "learning_rate": 0.00018195856991327136, + "loss": 0.5204, + "step": 17350 + }, + { + "epoch": 0.27079303675048355, + "grad_norm": 5.1940460205078125, + "learning_rate": 0.0001819481707951166, + "loss": 0.3476, + "step": 17360 + }, + { + "epoch": 0.27094902352280525, + "grad_norm": 1.314439296722412, + "learning_rate": 0.0001819377716769618, + "loss": 0.4811, + "step": 17370 + }, + { + "epoch": 0.27110501029512696, + "grad_norm": 0.738513171672821, + "learning_rate": 0.00018192737255880704, + "loss": 0.1572, + "step": 17380 + }, + { + "epoch": 0.27126099706744866, + "grad_norm": 1.8002641201019287, + "learning_rate": 0.00018191697344065223, + "loss": 0.263, + "step": 17390 + }, + { + "epoch": 0.27141698383977036, + "grad_norm": 0.9312915802001953, + "learning_rate": 0.00018190657432249748, + "loss": 0.3446, + "step": 17400 + }, + { + "epoch": 0.27157297061209207, + "grad_norm": 3.9765400886535645, + "learning_rate": 0.00018189617520434267, + "loss": 0.2543, + "step": 17410 + }, + { + "epoch": 0.2717289573844138, + "grad_norm": 1.4780479669570923, + "learning_rate": 0.00018188577608618792, + "loss": 0.107, + "step": 17420 + }, + { + "epoch": 0.27188494415673553, + "grad_norm": 0.39317771792411804, + "learning_rate": 0.0001818753769680331, + "loss": 0.5711, + "step": 17430 + }, + { + "epoch": 0.27204093092905723, + "grad_norm": 4.742455959320068, + "learning_rate": 0.00018186497784987836, + "loss": 0.3872, + "step": 17440 + }, + { + "epoch": 0.27219691770137894, + "grad_norm": 1.4516676664352417, + "learning_rate": 0.00018185457873172355, + "loss": 0.3826, + "step": 17450 + }, + { + "epoch": 0.27235290447370064, + "grad_norm": 2.7607524394989014, + "learning_rate": 0.0001818441796135688, + "loss": 0.2899, + "step": 17460 + }, + { + "epoch": 0.27250889124602234, + "grad_norm": 1.3420262336730957, + "learning_rate": 0.000181833780495414, + "loss": 0.0738, + "step": 17470 + }, + { + "epoch": 0.27266487801834405, + "grad_norm": 1.7739657163619995, + "learning_rate": 0.00018182338137725924, + "loss": 0.2175, + "step": 17480 + }, + { + "epoch": 0.27282086479066575, + "grad_norm": 1.1350417137145996, + "learning_rate": 0.00018181298225910443, + "loss": 0.3831, + "step": 17490 + }, + { + "epoch": 0.27297685156298745, + "grad_norm": 1.9588922262191772, + "learning_rate": 0.00018180258314094968, + "loss": 0.4804, + "step": 17500 + }, + { + "epoch": 0.27313283833530916, + "grad_norm": 0.41531097888946533, + "learning_rate": 0.00018179218402279487, + "loss": 0.2196, + "step": 17510 + }, + { + "epoch": 0.27328882510763086, + "grad_norm": 2.0884125232696533, + "learning_rate": 0.0001817817849046401, + "loss": 0.2074, + "step": 17520 + }, + { + "epoch": 0.27344481187995257, + "grad_norm": 0.015479146502912045, + "learning_rate": 0.0001817713857864853, + "loss": 0.2793, + "step": 17530 + }, + { + "epoch": 0.27360079865227427, + "grad_norm": 3.1188344955444336, + "learning_rate": 0.00018176098666833053, + "loss": 0.3488, + "step": 17540 + }, + { + "epoch": 0.27375678542459597, + "grad_norm": 1.54562246799469, + "learning_rate": 0.00018175058755017575, + "loss": 0.2574, + "step": 17550 + }, + { + "epoch": 0.2739127721969177, + "grad_norm": 0.850337564945221, + "learning_rate": 0.00018174018843202097, + "loss": 0.194, + "step": 17560 + }, + { + "epoch": 0.27406875896923943, + "grad_norm": 2.9046213626861572, + "learning_rate": 0.0001817297893138662, + "loss": 0.3307, + "step": 17570 + }, + { + "epoch": 0.27422474574156114, + "grad_norm": 2.5438475608825684, + "learning_rate": 0.0001817193901957114, + "loss": 0.2395, + "step": 17580 + }, + { + "epoch": 0.27438073251388284, + "grad_norm": 2.32556414604187, + "learning_rate": 0.00018170899107755663, + "loss": 0.5629, + "step": 17590 + }, + { + "epoch": 0.27453671928620454, + "grad_norm": 0.1941114366054535, + "learning_rate": 0.00018169859195940185, + "loss": 0.1768, + "step": 17600 + }, + { + "epoch": 0.27469270605852625, + "grad_norm": 1.0553919076919556, + "learning_rate": 0.00018168819284124707, + "loss": 0.2846, + "step": 17610 + }, + { + "epoch": 0.27484869283084795, + "grad_norm": 1.029302716255188, + "learning_rate": 0.00018167779372309229, + "loss": 0.3215, + "step": 17620 + }, + { + "epoch": 0.27500467960316965, + "grad_norm": 3.4507462978363037, + "learning_rate": 0.0001816673946049375, + "loss": 0.1537, + "step": 17630 + }, + { + "epoch": 0.27516066637549136, + "grad_norm": 6.206595420837402, + "learning_rate": 0.00018165699548678273, + "loss": 0.4816, + "step": 17640 + }, + { + "epoch": 0.27531665314781306, + "grad_norm": 1.1753982305526733, + "learning_rate": 0.00018164659636862794, + "loss": 0.5097, + "step": 17650 + }, + { + "epoch": 0.27547263992013477, + "grad_norm": 3.6227025985717773, + "learning_rate": 0.00018163619725047316, + "loss": 0.3409, + "step": 17660 + }, + { + "epoch": 0.27562862669245647, + "grad_norm": 0.20433004200458527, + "learning_rate": 0.00018162579813231838, + "loss": 0.1685, + "step": 17670 + }, + { + "epoch": 0.27578461346477817, + "grad_norm": 2.635998249053955, + "learning_rate": 0.0001816153990141636, + "loss": 0.2563, + "step": 17680 + }, + { + "epoch": 0.2759406002370999, + "grad_norm": 1.7162081003189087, + "learning_rate": 0.00018160499989600882, + "loss": 0.3069, + "step": 17690 + }, + { + "epoch": 0.2760965870094216, + "grad_norm": 2.9401514530181885, + "learning_rate": 0.00018159460077785404, + "loss": 0.3576, + "step": 17700 + }, + { + "epoch": 0.2762525737817433, + "grad_norm": 1.4552119970321655, + "learning_rate": 0.00018158420165969926, + "loss": 0.2456, + "step": 17710 + }, + { + "epoch": 0.27640856055406504, + "grad_norm": 2.02592396736145, + "learning_rate": 0.00018157380254154448, + "loss": 0.3204, + "step": 17720 + }, + { + "epoch": 0.27656454732638674, + "grad_norm": 0.28846657276153564, + "learning_rate": 0.0001815634034233897, + "loss": 0.1064, + "step": 17730 + }, + { + "epoch": 0.27672053409870845, + "grad_norm": 0.19064664840698242, + "learning_rate": 0.00018155300430523492, + "loss": 0.3128, + "step": 17740 + }, + { + "epoch": 0.27687652087103015, + "grad_norm": 0.9748614430427551, + "learning_rate": 0.00018154260518708014, + "loss": 0.5027, + "step": 17750 + }, + { + "epoch": 0.27703250764335186, + "grad_norm": 2.212595224380493, + "learning_rate": 0.00018153220606892536, + "loss": 0.181, + "step": 17760 + }, + { + "epoch": 0.27718849441567356, + "grad_norm": 1.0098037719726562, + "learning_rate": 0.00018152180695077058, + "loss": 0.3514, + "step": 17770 + }, + { + "epoch": 0.27734448118799526, + "grad_norm": 0.8791114091873169, + "learning_rate": 0.0001815114078326158, + "loss": 0.4152, + "step": 17780 + }, + { + "epoch": 0.27750046796031697, + "grad_norm": 1.3200514316558838, + "learning_rate": 0.00018150100871446102, + "loss": 0.3304, + "step": 17790 + }, + { + "epoch": 0.27765645473263867, + "grad_norm": 0.4182199239730835, + "learning_rate": 0.00018149060959630624, + "loss": 0.2791, + "step": 17800 + }, + { + "epoch": 0.2778124415049604, + "grad_norm": 0.9958666563034058, + "learning_rate": 0.00018148021047815146, + "loss": 0.3131, + "step": 17810 + }, + { + "epoch": 0.2779684282772821, + "grad_norm": 1.1399595737457275, + "learning_rate": 0.00018146981135999668, + "loss": 0.2365, + "step": 17820 + }, + { + "epoch": 0.2781244150496038, + "grad_norm": 5.128471851348877, + "learning_rate": 0.0001814594122418419, + "loss": 0.2908, + "step": 17830 + }, + { + "epoch": 0.2782804018219255, + "grad_norm": 4.693942070007324, + "learning_rate": 0.00018144901312368712, + "loss": 0.206, + "step": 17840 + }, + { + "epoch": 0.2784363885942472, + "grad_norm": 0.8024560809135437, + "learning_rate": 0.00018143861400553234, + "loss": 0.5016, + "step": 17850 + }, + { + "epoch": 0.2785923753665689, + "grad_norm": 1.2355642318725586, + "learning_rate": 0.00018142821488737756, + "loss": 0.1689, + "step": 17860 + }, + { + "epoch": 0.27874836213889065, + "grad_norm": 4.943911075592041, + "learning_rate": 0.00018141781576922278, + "loss": 0.3794, + "step": 17870 + }, + { + "epoch": 0.27890434891121235, + "grad_norm": 2.7630674839019775, + "learning_rate": 0.000181407416651068, + "loss": 0.3477, + "step": 17880 + }, + { + "epoch": 0.27906033568353406, + "grad_norm": 1.3990181684494019, + "learning_rate": 0.00018139701753291322, + "loss": 0.452, + "step": 17890 + }, + { + "epoch": 0.27921632245585576, + "grad_norm": 1.2540063858032227, + "learning_rate": 0.00018138661841475844, + "loss": 0.3852, + "step": 17900 + }, + { + "epoch": 0.27937230922817746, + "grad_norm": 4.327805995941162, + "learning_rate": 0.00018137621929660366, + "loss": 0.2202, + "step": 17910 + }, + { + "epoch": 0.27952829600049917, + "grad_norm": 2.9348056316375732, + "learning_rate": 0.00018136582017844888, + "loss": 0.2936, + "step": 17920 + }, + { + "epoch": 0.27968428277282087, + "grad_norm": 5.549226760864258, + "learning_rate": 0.0001813554210602941, + "loss": 0.3205, + "step": 17930 + }, + { + "epoch": 0.2798402695451426, + "grad_norm": 0.554421603679657, + "learning_rate": 0.00018134502194213931, + "loss": 0.2409, + "step": 17940 + }, + { + "epoch": 0.2799962563174643, + "grad_norm": 1.1400529146194458, + "learning_rate": 0.00018133462282398453, + "loss": 0.3497, + "step": 17950 + }, + { + "epoch": 0.280152243089786, + "grad_norm": 1.1388673782348633, + "learning_rate": 0.00018132422370582975, + "loss": 0.3676, + "step": 17960 + }, + { + "epoch": 0.2803082298621077, + "grad_norm": 1.0324344635009766, + "learning_rate": 0.00018131382458767497, + "loss": 0.3167, + "step": 17970 + }, + { + "epoch": 0.2804642166344294, + "grad_norm": 2.051694869995117, + "learning_rate": 0.0001813034254695202, + "loss": 0.2457, + "step": 17980 + }, + { + "epoch": 0.2806202034067511, + "grad_norm": 1.3411865234375, + "learning_rate": 0.0001812930263513654, + "loss": 0.1896, + "step": 17990 + }, + { + "epoch": 0.2807761901790728, + "grad_norm": 2.2051069736480713, + "learning_rate": 0.00018128262723321063, + "loss": 0.2585, + "step": 18000 + }, + { + "epoch": 0.2809321769513945, + "grad_norm": 0.4186108410358429, + "learning_rate": 0.00018127222811505585, + "loss": 0.1817, + "step": 18010 + }, + { + "epoch": 0.28108816372371626, + "grad_norm": 2.2891879081726074, + "learning_rate": 0.00018126182899690107, + "loss": 0.3138, + "step": 18020 + }, + { + "epoch": 0.28124415049603796, + "grad_norm": 1.0535072088241577, + "learning_rate": 0.0001812514298787463, + "loss": 0.5173, + "step": 18030 + }, + { + "epoch": 0.28140013726835966, + "grad_norm": 0.8798923492431641, + "learning_rate": 0.0001812410307605915, + "loss": 0.1406, + "step": 18040 + }, + { + "epoch": 0.28155612404068137, + "grad_norm": 1.1767810583114624, + "learning_rate": 0.00018123063164243673, + "loss": 0.3087, + "step": 18050 + }, + { + "epoch": 0.28171211081300307, + "grad_norm": 0.8934350609779358, + "learning_rate": 0.00018122023252428195, + "loss": 0.2565, + "step": 18060 + }, + { + "epoch": 0.2818680975853248, + "grad_norm": 0.58326655626297, + "learning_rate": 0.00018120983340612717, + "loss": 0.1834, + "step": 18070 + }, + { + "epoch": 0.2820240843576465, + "grad_norm": 2.871858835220337, + "learning_rate": 0.0001811994342879724, + "loss": 0.3183, + "step": 18080 + }, + { + "epoch": 0.2821800711299682, + "grad_norm": 2.103092670440674, + "learning_rate": 0.0001811890351698176, + "loss": 0.3188, + "step": 18090 + }, + { + "epoch": 0.2823360579022899, + "grad_norm": 0.4794228971004486, + "learning_rate": 0.00018117863605166283, + "loss": 0.239, + "step": 18100 + }, + { + "epoch": 0.2824920446746116, + "grad_norm": 1.2839410305023193, + "learning_rate": 0.00018116823693350805, + "loss": 0.3997, + "step": 18110 + }, + { + "epoch": 0.2826480314469333, + "grad_norm": 1.053501009941101, + "learning_rate": 0.00018115783781535327, + "loss": 0.2521, + "step": 18120 + }, + { + "epoch": 0.282804018219255, + "grad_norm": 2.066743850708008, + "learning_rate": 0.0001811474386971985, + "loss": 0.6073, + "step": 18130 + }, + { + "epoch": 0.2829600049915767, + "grad_norm": 3.7811508178710938, + "learning_rate": 0.0001811370395790437, + "loss": 0.3662, + "step": 18140 + }, + { + "epoch": 0.2831159917638984, + "grad_norm": 0.909883439540863, + "learning_rate": 0.00018112664046088893, + "loss": 0.227, + "step": 18150 + }, + { + "epoch": 0.2832719785362201, + "grad_norm": 0.4856087267398834, + "learning_rate": 0.00018111624134273415, + "loss": 0.2749, + "step": 18160 + }, + { + "epoch": 0.28342796530854186, + "grad_norm": 1.9020060300827026, + "learning_rate": 0.00018110584222457937, + "loss": 0.42, + "step": 18170 + }, + { + "epoch": 0.28358395208086357, + "grad_norm": 3.7515413761138916, + "learning_rate": 0.00018109544310642459, + "loss": 0.3674, + "step": 18180 + }, + { + "epoch": 0.28373993885318527, + "grad_norm": 0.06593445688486099, + "learning_rate": 0.0001810850439882698, + "loss": 0.2839, + "step": 18190 + }, + { + "epoch": 0.283895925625507, + "grad_norm": 1.6170812845230103, + "learning_rate": 0.00018107464487011503, + "loss": 0.3227, + "step": 18200 + }, + { + "epoch": 0.2840519123978287, + "grad_norm": 3.0923237800598145, + "learning_rate": 0.00018106424575196024, + "loss": 0.2593, + "step": 18210 + }, + { + "epoch": 0.2842078991701504, + "grad_norm": 5.011078357696533, + "learning_rate": 0.00018105384663380546, + "loss": 0.3599, + "step": 18220 + }, + { + "epoch": 0.2843638859424721, + "grad_norm": 0.6333115696907043, + "learning_rate": 0.00018104344751565068, + "loss": 0.2482, + "step": 18230 + }, + { + "epoch": 0.2845198727147938, + "grad_norm": 0.4594322443008423, + "learning_rate": 0.0001810330483974959, + "loss": 0.3561, + "step": 18240 + }, + { + "epoch": 0.2846758594871155, + "grad_norm": 3.7652647495269775, + "learning_rate": 0.00018102264927934112, + "loss": 0.3504, + "step": 18250 + }, + { + "epoch": 0.2848318462594372, + "grad_norm": 5.6666035652160645, + "learning_rate": 0.00018101225016118634, + "loss": 0.1973, + "step": 18260 + }, + { + "epoch": 0.2849878330317589, + "grad_norm": 0.4474485516548157, + "learning_rate": 0.00018100185104303156, + "loss": 0.1019, + "step": 18270 + }, + { + "epoch": 0.2851438198040806, + "grad_norm": 0.6686190962791443, + "learning_rate": 0.00018099145192487678, + "loss": 0.2669, + "step": 18280 + }, + { + "epoch": 0.2852998065764023, + "grad_norm": 3.2231101989746094, + "learning_rate": 0.000180981052806722, + "loss": 0.2856, + "step": 18290 + }, + { + "epoch": 0.285455793348724, + "grad_norm": 3.4004781246185303, + "learning_rate": 0.00018097065368856722, + "loss": 0.2206, + "step": 18300 + }, + { + "epoch": 0.2856117801210457, + "grad_norm": 0.14565567672252655, + "learning_rate": 0.00018096025457041244, + "loss": 0.2342, + "step": 18310 + }, + { + "epoch": 0.28576776689336747, + "grad_norm": 2.560617208480835, + "learning_rate": 0.00018094985545225766, + "loss": 0.497, + "step": 18320 + }, + { + "epoch": 0.2859237536656892, + "grad_norm": 1.615782618522644, + "learning_rate": 0.00018093945633410288, + "loss": 0.2626, + "step": 18330 + }, + { + "epoch": 0.2860797404380109, + "grad_norm": 0.10760047286748886, + "learning_rate": 0.0001809290572159481, + "loss": 0.1658, + "step": 18340 + }, + { + "epoch": 0.2862357272103326, + "grad_norm": 1.2961076498031616, + "learning_rate": 0.00018091865809779332, + "loss": 0.3492, + "step": 18350 + }, + { + "epoch": 0.2863917139826543, + "grad_norm": 2.04667067527771, + "learning_rate": 0.00018090825897963854, + "loss": 0.312, + "step": 18360 + }, + { + "epoch": 0.286547700754976, + "grad_norm": 1.5102049112319946, + "learning_rate": 0.00018089785986148376, + "loss": 0.2802, + "step": 18370 + }, + { + "epoch": 0.2867036875272977, + "grad_norm": 2.23567795753479, + "learning_rate": 0.00018088746074332898, + "loss": 0.3466, + "step": 18380 + }, + { + "epoch": 0.2868596742996194, + "grad_norm": 3.9257795810699463, + "learning_rate": 0.0001808770616251742, + "loss": 0.4444, + "step": 18390 + }, + { + "epoch": 0.2870156610719411, + "grad_norm": 0.5947315096855164, + "learning_rate": 0.00018086666250701942, + "loss": 0.1304, + "step": 18400 + }, + { + "epoch": 0.2871716478442628, + "grad_norm": 8.322796821594238, + "learning_rate": 0.00018085626338886464, + "loss": 0.2911, + "step": 18410 + }, + { + "epoch": 0.2873276346165845, + "grad_norm": 0.14317284524440765, + "learning_rate": 0.00018084586427070983, + "loss": 0.1981, + "step": 18420 + }, + { + "epoch": 0.2874836213889062, + "grad_norm": 2.292736291885376, + "learning_rate": 0.00018083546515255508, + "loss": 0.2705, + "step": 18430 + }, + { + "epoch": 0.2876396081612279, + "grad_norm": 2.2122421264648438, + "learning_rate": 0.00018082506603440027, + "loss": 0.2954, + "step": 18440 + }, + { + "epoch": 0.2877955949335496, + "grad_norm": 4.216709613800049, + "learning_rate": 0.00018081466691624552, + "loss": 0.3587, + "step": 18450 + }, + { + "epoch": 0.2879515817058713, + "grad_norm": 1.9166659116744995, + "learning_rate": 0.0001808042677980907, + "loss": 0.2069, + "step": 18460 + }, + { + "epoch": 0.2881075684781931, + "grad_norm": 1.233359694480896, + "learning_rate": 0.00018079386867993596, + "loss": 0.14, + "step": 18470 + }, + { + "epoch": 0.2882635552505148, + "grad_norm": 3.25506591796875, + "learning_rate": 0.00018078346956178115, + "loss": 0.2339, + "step": 18480 + }, + { + "epoch": 0.2884195420228365, + "grad_norm": 5.1549272537231445, + "learning_rate": 0.0001807730704436264, + "loss": 0.409, + "step": 18490 + }, + { + "epoch": 0.2885755287951582, + "grad_norm": 4.583444118499756, + "learning_rate": 0.0001807626713254716, + "loss": 0.3163, + "step": 18500 + }, + { + "epoch": 0.2887315155674799, + "grad_norm": 1.3562095165252686, + "learning_rate": 0.00018075227220731683, + "loss": 0.2213, + "step": 18510 + }, + { + "epoch": 0.2888875023398016, + "grad_norm": 0.674741268157959, + "learning_rate": 0.00018074187308916203, + "loss": 0.2659, + "step": 18520 + }, + { + "epoch": 0.2890434891121233, + "grad_norm": 0.1102909967303276, + "learning_rate": 0.00018073147397100727, + "loss": 0.2049, + "step": 18530 + }, + { + "epoch": 0.289199475884445, + "grad_norm": 0.121336430311203, + "learning_rate": 0.00018072107485285247, + "loss": 0.0701, + "step": 18540 + }, + { + "epoch": 0.2893554626567667, + "grad_norm": 0.7439133524894714, + "learning_rate": 0.0001807106757346977, + "loss": 0.1598, + "step": 18550 + }, + { + "epoch": 0.2895114494290884, + "grad_norm": 1.0136058330535889, + "learning_rate": 0.0001807002766165429, + "loss": 0.3005, + "step": 18560 + }, + { + "epoch": 0.2896674362014101, + "grad_norm": 2.1980068683624268, + "learning_rate": 0.00018068987749838815, + "loss": 0.2871, + "step": 18570 + }, + { + "epoch": 0.2898234229737318, + "grad_norm": 2.9443399906158447, + "learning_rate": 0.00018067947838023337, + "loss": 0.336, + "step": 18580 + }, + { + "epoch": 0.2899794097460535, + "grad_norm": 0.9509536623954773, + "learning_rate": 0.0001806690792620786, + "loss": 0.2991, + "step": 18590 + }, + { + "epoch": 0.2901353965183752, + "grad_norm": 1.0886099338531494, + "learning_rate": 0.0001806586801439238, + "loss": 0.1968, + "step": 18600 + }, + { + "epoch": 0.2902913832906969, + "grad_norm": 2.102788209915161, + "learning_rate": 0.00018064828102576903, + "loss": 0.2603, + "step": 18610 + }, + { + "epoch": 0.29044737006301863, + "grad_norm": 1.910498023033142, + "learning_rate": 0.00018063788190761425, + "loss": 0.5918, + "step": 18620 + }, + { + "epoch": 0.2906033568353404, + "grad_norm": 2.6424412727355957, + "learning_rate": 0.00018062748278945947, + "loss": 0.28, + "step": 18630 + }, + { + "epoch": 0.2907593436076621, + "grad_norm": 1.333108901977539, + "learning_rate": 0.0001806170836713047, + "loss": 0.3092, + "step": 18640 + }, + { + "epoch": 0.2909153303799838, + "grad_norm": 2.399016857147217, + "learning_rate": 0.0001806066845531499, + "loss": 0.1958, + "step": 18650 + }, + { + "epoch": 0.2910713171523055, + "grad_norm": 2.875866413116455, + "learning_rate": 0.00018059628543499513, + "loss": 0.2086, + "step": 18660 + }, + { + "epoch": 0.2912273039246272, + "grad_norm": 1.014180064201355, + "learning_rate": 0.00018058588631684035, + "loss": 0.2348, + "step": 18670 + }, + { + "epoch": 0.2913832906969489, + "grad_norm": 2.1878981590270996, + "learning_rate": 0.00018057548719868557, + "loss": 0.3789, + "step": 18680 + }, + { + "epoch": 0.2915392774692706, + "grad_norm": 1.0153847932815552, + "learning_rate": 0.0001805650880805308, + "loss": 0.1895, + "step": 18690 + }, + { + "epoch": 0.2916952642415923, + "grad_norm": 3.4147262573242188, + "learning_rate": 0.000180554688962376, + "loss": 0.2737, + "step": 18700 + }, + { + "epoch": 0.291851251013914, + "grad_norm": 1.0009074211120605, + "learning_rate": 0.00018054428984422123, + "loss": 0.2325, + "step": 18710 + }, + { + "epoch": 0.2920072377862357, + "grad_norm": 0.05532321706414223, + "learning_rate": 0.00018053389072606645, + "loss": 0.3217, + "step": 18720 + }, + { + "epoch": 0.2921632245585574, + "grad_norm": 1.0321491956710815, + "learning_rate": 0.00018052349160791167, + "loss": 0.2416, + "step": 18730 + }, + { + "epoch": 0.2923192113308791, + "grad_norm": 1.7075380086898804, + "learning_rate": 0.00018051309248975689, + "loss": 0.4041, + "step": 18740 + }, + { + "epoch": 0.29247519810320083, + "grad_norm": 3.3456871509552, + "learning_rate": 0.0001805026933716021, + "loss": 0.2507, + "step": 18750 + }, + { + "epoch": 0.29263118487552253, + "grad_norm": 1.829784631729126, + "learning_rate": 0.00018049229425344732, + "loss": 0.1845, + "step": 18760 + }, + { + "epoch": 0.29278717164784424, + "grad_norm": 0.0910114273428917, + "learning_rate": 0.00018048189513529254, + "loss": 0.1832, + "step": 18770 + }, + { + "epoch": 0.292943158420166, + "grad_norm": 1.5118619203567505, + "learning_rate": 0.00018047149601713776, + "loss": 0.2895, + "step": 18780 + }, + { + "epoch": 0.2930991451924877, + "grad_norm": 0.6485610008239746, + "learning_rate": 0.00018046109689898298, + "loss": 0.4677, + "step": 18790 + }, + { + "epoch": 0.2932551319648094, + "grad_norm": 2.6253087520599365, + "learning_rate": 0.0001804506977808282, + "loss": 0.249, + "step": 18800 + }, + { + "epoch": 0.2934111187371311, + "grad_norm": 2.591935396194458, + "learning_rate": 0.00018044029866267342, + "loss": 0.3311, + "step": 18810 + }, + { + "epoch": 0.2935671055094528, + "grad_norm": 1.5376029014587402, + "learning_rate": 0.00018042989954451864, + "loss": 0.2681, + "step": 18820 + }, + { + "epoch": 0.2937230922817745, + "grad_norm": 0.39641091227531433, + "learning_rate": 0.00018041950042636386, + "loss": 0.1739, + "step": 18830 + }, + { + "epoch": 0.2938790790540962, + "grad_norm": 0.6871474981307983, + "learning_rate": 0.00018040910130820908, + "loss": 0.2747, + "step": 18840 + }, + { + "epoch": 0.2940350658264179, + "grad_norm": 0.7258870005607605, + "learning_rate": 0.0001803987021900543, + "loss": 0.2717, + "step": 18850 + }, + { + "epoch": 0.2941910525987396, + "grad_norm": 2.1410534381866455, + "learning_rate": 0.00018038830307189952, + "loss": 0.3864, + "step": 18860 + }, + { + "epoch": 0.2943470393710613, + "grad_norm": 1.0177247524261475, + "learning_rate": 0.00018037790395374471, + "loss": 0.2783, + "step": 18870 + }, + { + "epoch": 0.29450302614338303, + "grad_norm": 0.9840215444564819, + "learning_rate": 0.00018036750483558996, + "loss": 0.2108, + "step": 18880 + }, + { + "epoch": 0.29465901291570473, + "grad_norm": 6.390523910522461, + "learning_rate": 0.00018035710571743515, + "loss": 0.3335, + "step": 18890 + }, + { + "epoch": 0.29481499968802644, + "grad_norm": 1.692339539527893, + "learning_rate": 0.0001803467065992804, + "loss": 0.3347, + "step": 18900 + }, + { + "epoch": 0.29497098646034814, + "grad_norm": 1.4099748134613037, + "learning_rate": 0.0001803363074811256, + "loss": 0.3066, + "step": 18910 + }, + { + "epoch": 0.29512697323266984, + "grad_norm": 0.43823370337486267, + "learning_rate": 0.00018032590836297084, + "loss": 0.3014, + "step": 18920 + }, + { + "epoch": 0.2952829600049916, + "grad_norm": 2.873828411102295, + "learning_rate": 0.00018031550924481603, + "loss": 0.1334, + "step": 18930 + }, + { + "epoch": 0.2954389467773133, + "grad_norm": 1.8268390893936157, + "learning_rate": 0.00018030511012666128, + "loss": 0.1313, + "step": 18940 + }, + { + "epoch": 0.295594933549635, + "grad_norm": 1.147981882095337, + "learning_rate": 0.00018029471100850647, + "loss": 0.1788, + "step": 18950 + }, + { + "epoch": 0.2957509203219567, + "grad_norm": 0.8453661799430847, + "learning_rate": 0.00018028431189035172, + "loss": 0.3169, + "step": 18960 + }, + { + "epoch": 0.2959069070942784, + "grad_norm": 1.97072172164917, + "learning_rate": 0.0001802739127721969, + "loss": 0.3351, + "step": 18970 + }, + { + "epoch": 0.2960628938666001, + "grad_norm": 3.8672759532928467, + "learning_rate": 0.00018026351365404216, + "loss": 0.5419, + "step": 18980 + }, + { + "epoch": 0.2962188806389218, + "grad_norm": 0.5253915190696716, + "learning_rate": 0.00018025311453588735, + "loss": 0.3463, + "step": 18990 + }, + { + "epoch": 0.2963748674112435, + "grad_norm": 0.48347389698028564, + "learning_rate": 0.0001802427154177326, + "loss": 0.3805, + "step": 19000 + }, + { + "epoch": 0.29653085418356523, + "grad_norm": 1.1975009441375732, + "learning_rate": 0.0001802323162995778, + "loss": 0.1402, + "step": 19010 + }, + { + "epoch": 0.29668684095588693, + "grad_norm": 1.5407238006591797, + "learning_rate": 0.00018022191718142304, + "loss": 0.215, + "step": 19020 + }, + { + "epoch": 0.29684282772820864, + "grad_norm": 0.4473857581615448, + "learning_rate": 0.00018021151806326823, + "loss": 0.2989, + "step": 19030 + }, + { + "epoch": 0.29699881450053034, + "grad_norm": 2.9506404399871826, + "learning_rate": 0.00018020111894511347, + "loss": 0.2944, + "step": 19040 + }, + { + "epoch": 0.29715480127285204, + "grad_norm": 2.249565601348877, + "learning_rate": 0.00018019071982695867, + "loss": 0.1829, + "step": 19050 + }, + { + "epoch": 0.29731078804517375, + "grad_norm": 2.1028025150299072, + "learning_rate": 0.00018018032070880391, + "loss": 0.2226, + "step": 19060 + }, + { + "epoch": 0.29746677481749545, + "grad_norm": 3.834991216659546, + "learning_rate": 0.0001801699215906491, + "loss": 0.5616, + "step": 19070 + }, + { + "epoch": 0.2976227615898172, + "grad_norm": 1.1924679279327393, + "learning_rate": 0.00018015952247249435, + "loss": 0.3003, + "step": 19080 + }, + { + "epoch": 0.2977787483621389, + "grad_norm": 0.6356360912322998, + "learning_rate": 0.00018014912335433955, + "loss": 0.1035, + "step": 19090 + }, + { + "epoch": 0.2979347351344606, + "grad_norm": 0.015022732317447662, + "learning_rate": 0.0001801387242361848, + "loss": 0.1645, + "step": 19100 + }, + { + "epoch": 0.2980907219067823, + "grad_norm": 1.1007914543151855, + "learning_rate": 0.00018012832511802999, + "loss": 0.1699, + "step": 19110 + }, + { + "epoch": 0.298246708679104, + "grad_norm": 1.9343289136886597, + "learning_rate": 0.00018011792599987523, + "loss": 0.4052, + "step": 19120 + }, + { + "epoch": 0.2984026954514257, + "grad_norm": 0.6810970306396484, + "learning_rate": 0.00018010752688172042, + "loss": 0.4275, + "step": 19130 + }, + { + "epoch": 0.29855868222374743, + "grad_norm": 0.9088418483734131, + "learning_rate": 0.00018009712776356567, + "loss": 0.3932, + "step": 19140 + }, + { + "epoch": 0.29871466899606913, + "grad_norm": 1.6194713115692139, + "learning_rate": 0.00018008672864541086, + "loss": 0.2688, + "step": 19150 + }, + { + "epoch": 0.29887065576839084, + "grad_norm": 3.2220847606658936, + "learning_rate": 0.0001800763295272561, + "loss": 0.3467, + "step": 19160 + }, + { + "epoch": 0.29902664254071254, + "grad_norm": 2.351975679397583, + "learning_rate": 0.0001800659304091013, + "loss": 0.4058, + "step": 19170 + }, + { + "epoch": 0.29918262931303424, + "grad_norm": 2.682670831680298, + "learning_rate": 0.00018005553129094655, + "loss": 0.1584, + "step": 19180 + }, + { + "epoch": 0.29933861608535595, + "grad_norm": 2.738229513168335, + "learning_rate": 0.00018004513217279174, + "loss": 0.3897, + "step": 19190 + }, + { + "epoch": 0.29949460285767765, + "grad_norm": 0.952240526676178, + "learning_rate": 0.000180034733054637, + "loss": 0.368, + "step": 19200 + }, + { + "epoch": 0.29965058962999935, + "grad_norm": 2.784024238586426, + "learning_rate": 0.00018002433393648218, + "loss": 0.3375, + "step": 19210 + }, + { + "epoch": 0.29980657640232106, + "grad_norm": 0.45826661586761475, + "learning_rate": 0.00018001393481832743, + "loss": 0.1703, + "step": 19220 + }, + { + "epoch": 0.2999625631746428, + "grad_norm": 3.0322210788726807, + "learning_rate": 0.00018000353570017262, + "loss": 0.3613, + "step": 19230 + }, + { + "epoch": 0.3001185499469645, + "grad_norm": 1.0897084474563599, + "learning_rate": 0.00017999313658201787, + "loss": 0.4094, + "step": 19240 + }, + { + "epoch": 0.3002745367192862, + "grad_norm": 1.2861119508743286, + "learning_rate": 0.00017998273746386306, + "loss": 0.1848, + "step": 19250 + }, + { + "epoch": 0.3004305234916079, + "grad_norm": 1.6010128259658813, + "learning_rate": 0.0001799723383457083, + "loss": 0.3557, + "step": 19260 + }, + { + "epoch": 0.30058651026392963, + "grad_norm": 1.2477325201034546, + "learning_rate": 0.0001799619392275535, + "loss": 0.4457, + "step": 19270 + }, + { + "epoch": 0.30074249703625133, + "grad_norm": 6.801977634429932, + "learning_rate": 0.00017995154010939875, + "loss": 0.1964, + "step": 19280 + }, + { + "epoch": 0.30089848380857304, + "grad_norm": 0.5982523560523987, + "learning_rate": 0.00017994114099124394, + "loss": 0.2198, + "step": 19290 + }, + { + "epoch": 0.30105447058089474, + "grad_norm": 0.46550920605659485, + "learning_rate": 0.00017993074187308919, + "loss": 0.0819, + "step": 19300 + }, + { + "epoch": 0.30121045735321644, + "grad_norm": 0.3265513479709625, + "learning_rate": 0.00017992034275493438, + "loss": 0.1301, + "step": 19310 + }, + { + "epoch": 0.30136644412553815, + "grad_norm": 0.4576639235019684, + "learning_rate": 0.0001799099436367796, + "loss": 0.2051, + "step": 19320 + }, + { + "epoch": 0.30152243089785985, + "grad_norm": 3.502267837524414, + "learning_rate": 0.00017989954451862482, + "loss": 0.393, + "step": 19330 + }, + { + "epoch": 0.30167841767018155, + "grad_norm": 2.840001344680786, + "learning_rate": 0.00017988914540047004, + "loss": 0.4381, + "step": 19340 + }, + { + "epoch": 0.30183440444250326, + "grad_norm": 1.0464013814926147, + "learning_rate": 0.00017987874628231526, + "loss": 0.217, + "step": 19350 + }, + { + "epoch": 0.30199039121482496, + "grad_norm": 3.0963642597198486, + "learning_rate": 0.00017986834716416048, + "loss": 0.2446, + "step": 19360 + }, + { + "epoch": 0.30214637798714666, + "grad_norm": 0.2639450430870056, + "learning_rate": 0.0001798579480460057, + "loss": 0.2497, + "step": 19370 + }, + { + "epoch": 0.3023023647594684, + "grad_norm": 2.2987661361694336, + "learning_rate": 0.00017984754892785092, + "loss": 0.428, + "step": 19380 + }, + { + "epoch": 0.3024583515317901, + "grad_norm": 0.41670629382133484, + "learning_rate": 0.00017983714980969614, + "loss": 0.1746, + "step": 19390 + }, + { + "epoch": 0.30261433830411183, + "grad_norm": 2.1702284812927246, + "learning_rate": 0.00017982675069154135, + "loss": 0.2832, + "step": 19400 + }, + { + "epoch": 0.30277032507643353, + "grad_norm": 0.51155024766922, + "learning_rate": 0.00017981635157338657, + "loss": 0.2777, + "step": 19410 + }, + { + "epoch": 0.30292631184875524, + "grad_norm": 1.4663769006729126, + "learning_rate": 0.0001798059524552318, + "loss": 0.4584, + "step": 19420 + }, + { + "epoch": 0.30308229862107694, + "grad_norm": 1.5726076364517212, + "learning_rate": 0.00017979555333707704, + "loss": 0.2313, + "step": 19430 + }, + { + "epoch": 0.30323828539339864, + "grad_norm": 1.7691172361373901, + "learning_rate": 0.00017978515421892223, + "loss": 0.3821, + "step": 19440 + }, + { + "epoch": 0.30339427216572035, + "grad_norm": 1.8396515846252441, + "learning_rate": 0.00017977475510076748, + "loss": 0.3427, + "step": 19450 + }, + { + "epoch": 0.30355025893804205, + "grad_norm": 3.1816394329071045, + "learning_rate": 0.00017976435598261267, + "loss": 0.3822, + "step": 19460 + }, + { + "epoch": 0.30370624571036375, + "grad_norm": 1.9252829551696777, + "learning_rate": 0.00017975395686445792, + "loss": 0.3783, + "step": 19470 + }, + { + "epoch": 0.30386223248268546, + "grad_norm": 1.5809651613235474, + "learning_rate": 0.0001797435577463031, + "loss": 0.4197, + "step": 19480 + }, + { + "epoch": 0.30401821925500716, + "grad_norm": 0.8304970860481262, + "learning_rate": 0.00017973315862814836, + "loss": 0.2594, + "step": 19490 + }, + { + "epoch": 0.30417420602732886, + "grad_norm": 0.6799660921096802, + "learning_rate": 0.00017972275950999355, + "loss": 0.1469, + "step": 19500 + }, + { + "epoch": 0.30433019279965057, + "grad_norm": 0.7135075926780701, + "learning_rate": 0.0001797123603918388, + "loss": 0.2243, + "step": 19510 + }, + { + "epoch": 0.30448617957197227, + "grad_norm": 0.748047411441803, + "learning_rate": 0.000179701961273684, + "loss": 0.1626, + "step": 19520 + }, + { + "epoch": 0.30464216634429403, + "grad_norm": 1.0538876056671143, + "learning_rate": 0.00017969156215552924, + "loss": 0.3698, + "step": 19530 + }, + { + "epoch": 0.30479815311661573, + "grad_norm": 2.149568796157837, + "learning_rate": 0.00017968116303737443, + "loss": 0.3964, + "step": 19540 + }, + { + "epoch": 0.30495413988893744, + "grad_norm": 6.2351179122924805, + "learning_rate": 0.00017967076391921968, + "loss": 0.3108, + "step": 19550 + }, + { + "epoch": 0.30511012666125914, + "grad_norm": 1.6175888776779175, + "learning_rate": 0.00017966036480106487, + "loss": 0.3445, + "step": 19560 + }, + { + "epoch": 0.30526611343358084, + "grad_norm": 1.9418601989746094, + "learning_rate": 0.00017964996568291012, + "loss": 0.1622, + "step": 19570 + }, + { + "epoch": 0.30542210020590255, + "grad_norm": 1.170223593711853, + "learning_rate": 0.0001796395665647553, + "loss": 0.2429, + "step": 19580 + }, + { + "epoch": 0.30557808697822425, + "grad_norm": 0.35105133056640625, + "learning_rate": 0.00017962916744660056, + "loss": 0.2293, + "step": 19590 + }, + { + "epoch": 0.30573407375054595, + "grad_norm": 1.9052326679229736, + "learning_rate": 0.00017961876832844575, + "loss": 0.2934, + "step": 19600 + }, + { + "epoch": 0.30589006052286766, + "grad_norm": 1.8728259801864624, + "learning_rate": 0.000179608369210291, + "loss": 0.3007, + "step": 19610 + }, + { + "epoch": 0.30604604729518936, + "grad_norm": 2.135467290878296, + "learning_rate": 0.0001795979700921362, + "loss": 0.3738, + "step": 19620 + }, + { + "epoch": 0.30620203406751106, + "grad_norm": 0.4691134989261627, + "learning_rate": 0.00017958757097398143, + "loss": 0.3777, + "step": 19630 + }, + { + "epoch": 0.30635802083983277, + "grad_norm": 1.0836161375045776, + "learning_rate": 0.00017957717185582663, + "loss": 0.2855, + "step": 19640 + }, + { + "epoch": 0.30651400761215447, + "grad_norm": 1.7246626615524292, + "learning_rate": 0.00017956677273767187, + "loss": 0.5668, + "step": 19650 + }, + { + "epoch": 0.3066699943844762, + "grad_norm": 0.40280747413635254, + "learning_rate": 0.00017955637361951707, + "loss": 0.0941, + "step": 19660 + }, + { + "epoch": 0.3068259811567979, + "grad_norm": 1.5181760787963867, + "learning_rate": 0.0001795459745013623, + "loss": 0.3913, + "step": 19670 + }, + { + "epoch": 0.30698196792911964, + "grad_norm": 0.9640255570411682, + "learning_rate": 0.0001795355753832075, + "loss": 0.3722, + "step": 19680 + }, + { + "epoch": 0.30713795470144134, + "grad_norm": 1.673332691192627, + "learning_rate": 0.00017952517626505275, + "loss": 0.2569, + "step": 19690 + }, + { + "epoch": 0.30729394147376304, + "grad_norm": 0.7451714873313904, + "learning_rate": 0.00017951477714689794, + "loss": 0.3577, + "step": 19700 + }, + { + "epoch": 0.30744992824608475, + "grad_norm": 2.918775796890259, + "learning_rate": 0.0001795043780287432, + "loss": 0.2836, + "step": 19710 + }, + { + "epoch": 0.30760591501840645, + "grad_norm": 1.9705866575241089, + "learning_rate": 0.00017949397891058838, + "loss": 0.2471, + "step": 19720 + }, + { + "epoch": 0.30776190179072815, + "grad_norm": 0.7021917104721069, + "learning_rate": 0.00017948357979243363, + "loss": 0.1858, + "step": 19730 + }, + { + "epoch": 0.30791788856304986, + "grad_norm": 1.7279014587402344, + "learning_rate": 0.00017947318067427882, + "loss": 0.2495, + "step": 19740 + }, + { + "epoch": 0.30807387533537156, + "grad_norm": 1.9975436925888062, + "learning_rate": 0.00017946278155612407, + "loss": 0.2416, + "step": 19750 + }, + { + "epoch": 0.30822986210769326, + "grad_norm": 3.755798816680908, + "learning_rate": 0.00017945238243796926, + "loss": 0.2197, + "step": 19760 + }, + { + "epoch": 0.30838584888001497, + "grad_norm": 7.889193058013916, + "learning_rate": 0.00017944198331981448, + "loss": 0.2867, + "step": 19770 + }, + { + "epoch": 0.30854183565233667, + "grad_norm": 0.42436349391937256, + "learning_rate": 0.0001794315842016597, + "loss": 0.3439, + "step": 19780 + }, + { + "epoch": 0.3086978224246584, + "grad_norm": 1.4928020238876343, + "learning_rate": 0.00017942118508350492, + "loss": 0.1206, + "step": 19790 + }, + { + "epoch": 0.3088538091969801, + "grad_norm": 2.7025113105773926, + "learning_rate": 0.00017941078596535014, + "loss": 0.2189, + "step": 19800 + }, + { + "epoch": 0.3090097959693018, + "grad_norm": 5.512676239013672, + "learning_rate": 0.00017940038684719536, + "loss": 0.1156, + "step": 19810 + }, + { + "epoch": 0.3091657827416235, + "grad_norm": 0.9146280884742737, + "learning_rate": 0.00017938998772904058, + "loss": 0.3538, + "step": 19820 + }, + { + "epoch": 0.30932176951394524, + "grad_norm": 5.226728916168213, + "learning_rate": 0.0001793795886108858, + "loss": 0.2686, + "step": 19830 + }, + { + "epoch": 0.30947775628626695, + "grad_norm": 0.9321868419647217, + "learning_rate": 0.00017936918949273102, + "loss": 0.1891, + "step": 19840 + }, + { + "epoch": 0.30963374305858865, + "grad_norm": 1.2203339338302612, + "learning_rate": 0.00017935879037457624, + "loss": 0.2242, + "step": 19850 + }, + { + "epoch": 0.30978972983091035, + "grad_norm": 0.05168648064136505, + "learning_rate": 0.00017934839125642146, + "loss": 0.3073, + "step": 19860 + }, + { + "epoch": 0.30994571660323206, + "grad_norm": 2.3573200702667236, + "learning_rate": 0.00017933799213826668, + "loss": 0.2863, + "step": 19870 + }, + { + "epoch": 0.31010170337555376, + "grad_norm": 0.39311453700065613, + "learning_rate": 0.0001793275930201119, + "loss": 0.1939, + "step": 19880 + }, + { + "epoch": 0.31025769014787546, + "grad_norm": 2.5007941722869873, + "learning_rate": 0.00017931719390195712, + "loss": 0.2084, + "step": 19890 + }, + { + "epoch": 0.31041367692019717, + "grad_norm": 0.272157222032547, + "learning_rate": 0.00017930679478380234, + "loss": 0.2186, + "step": 19900 + }, + { + "epoch": 0.31056966369251887, + "grad_norm": 0.4767307639122009, + "learning_rate": 0.00017929639566564756, + "loss": 0.3897, + "step": 19910 + }, + { + "epoch": 0.3107256504648406, + "grad_norm": 1.7147952318191528, + "learning_rate": 0.00017928599654749278, + "loss": 0.0832, + "step": 19920 + }, + { + "epoch": 0.3108816372371623, + "grad_norm": 1.539294958114624, + "learning_rate": 0.000179275597429338, + "loss": 0.1433, + "step": 19930 + }, + { + "epoch": 0.311037624009484, + "grad_norm": 3.3402135372161865, + "learning_rate": 0.00017926519831118322, + "loss": 0.4807, + "step": 19940 + }, + { + "epoch": 0.3111936107818057, + "grad_norm": 1.8989461660385132, + "learning_rate": 0.00017925479919302844, + "loss": 0.1565, + "step": 19950 + }, + { + "epoch": 0.3113495975541274, + "grad_norm": 0.7807097434997559, + "learning_rate": 0.00017924440007487365, + "loss": 0.3372, + "step": 19960 + }, + { + "epoch": 0.3115055843264491, + "grad_norm": 2.9375503063201904, + "learning_rate": 0.00017923400095671887, + "loss": 0.4074, + "step": 19970 + }, + { + "epoch": 0.31166157109877085, + "grad_norm": 21.00003433227539, + "learning_rate": 0.0001792236018385641, + "loss": 0.375, + "step": 19980 + }, + { + "epoch": 0.31181755787109255, + "grad_norm": 2.188424825668335, + "learning_rate": 0.00017921320272040931, + "loss": 0.3877, + "step": 19990 + }, + { + "epoch": 0.31197354464341426, + "grad_norm": 0.4043445587158203, + "learning_rate": 0.00017920280360225453, + "loss": 0.2248, + "step": 20000 + }, + { + "epoch": 0.31212953141573596, + "grad_norm": 2.4816536903381348, + "learning_rate": 0.00017919240448409975, + "loss": 0.2999, + "step": 20010 + }, + { + "epoch": 0.31228551818805766, + "grad_norm": 1.9566909074783325, + "learning_rate": 0.00017918200536594497, + "loss": 0.3409, + "step": 20020 + }, + { + "epoch": 0.31244150496037937, + "grad_norm": 0.5148131847381592, + "learning_rate": 0.0001791716062477902, + "loss": 0.3496, + "step": 20030 + }, + { + "epoch": 0.31259749173270107, + "grad_norm": 2.177743434906006, + "learning_rate": 0.0001791612071296354, + "loss": 0.2678, + "step": 20040 + }, + { + "epoch": 0.3127534785050228, + "grad_norm": 1.1060746908187866, + "learning_rate": 0.00017915080801148063, + "loss": 0.2213, + "step": 20050 + }, + { + "epoch": 0.3129094652773445, + "grad_norm": 0.3918687105178833, + "learning_rate": 0.00017914040889332585, + "loss": 0.3218, + "step": 20060 + }, + { + "epoch": 0.3130654520496662, + "grad_norm": 2.1818249225616455, + "learning_rate": 0.00017913000977517107, + "loss": 0.2117, + "step": 20070 + }, + { + "epoch": 0.3132214388219879, + "grad_norm": 0.5774380564689636, + "learning_rate": 0.0001791196106570163, + "loss": 0.1693, + "step": 20080 + }, + { + "epoch": 0.3133774255943096, + "grad_norm": 2.403837203979492, + "learning_rate": 0.0001791092115388615, + "loss": 0.3325, + "step": 20090 + }, + { + "epoch": 0.3135334123666313, + "grad_norm": 1.862305998802185, + "learning_rate": 0.00017909881242070673, + "loss": 0.185, + "step": 20100 + }, + { + "epoch": 0.313689399138953, + "grad_norm": 2.9921820163726807, + "learning_rate": 0.00017908841330255195, + "loss": 0.3907, + "step": 20110 + }, + { + "epoch": 0.3138453859112747, + "grad_norm": 1.529916763305664, + "learning_rate": 0.00017907801418439717, + "loss": 0.2994, + "step": 20120 + }, + { + "epoch": 0.31400137268359646, + "grad_norm": 0.1132139042019844, + "learning_rate": 0.0001790676150662424, + "loss": 0.1278, + "step": 20130 + }, + { + "epoch": 0.31415735945591816, + "grad_norm": 2.252502918243408, + "learning_rate": 0.0001790572159480876, + "loss": 0.2654, + "step": 20140 + }, + { + "epoch": 0.31431334622823986, + "grad_norm": 1.7733968496322632, + "learning_rate": 0.00017904681682993283, + "loss": 0.2088, + "step": 20150 + }, + { + "epoch": 0.31446933300056157, + "grad_norm": 0.07410364598035812, + "learning_rate": 0.00017903641771177805, + "loss": 0.3582, + "step": 20160 + }, + { + "epoch": 0.31462531977288327, + "grad_norm": 0.7408519387245178, + "learning_rate": 0.00017902601859362327, + "loss": 0.2341, + "step": 20170 + }, + { + "epoch": 0.314781306545205, + "grad_norm": 1.7559107542037964, + "learning_rate": 0.0001790156194754685, + "loss": 0.1483, + "step": 20180 + }, + { + "epoch": 0.3149372933175267, + "grad_norm": 1.763396143913269, + "learning_rate": 0.0001790052203573137, + "loss": 0.2536, + "step": 20190 + }, + { + "epoch": 0.3150932800898484, + "grad_norm": 0.82420414686203, + "learning_rate": 0.00017899482123915893, + "loss": 0.4354, + "step": 20200 + }, + { + "epoch": 0.3152492668621701, + "grad_norm": 2.74735689163208, + "learning_rate": 0.00017898442212100415, + "loss": 0.3735, + "step": 20210 + }, + { + "epoch": 0.3154052536344918, + "grad_norm": 3.1632747650146484, + "learning_rate": 0.00017897402300284937, + "loss": 0.1907, + "step": 20220 + }, + { + "epoch": 0.3155612404068135, + "grad_norm": 1.91727614402771, + "learning_rate": 0.00017896362388469459, + "loss": 0.2063, + "step": 20230 + }, + { + "epoch": 0.3157172271791352, + "grad_norm": 0.1301402896642685, + "learning_rate": 0.0001789532247665398, + "loss": 0.2402, + "step": 20240 + }, + { + "epoch": 0.3158732139514569, + "grad_norm": 0.243203803896904, + "learning_rate": 0.00017894282564838502, + "loss": 0.4258, + "step": 20250 + }, + { + "epoch": 0.3160292007237786, + "grad_norm": 1.6435601711273193, + "learning_rate": 0.00017893242653023024, + "loss": 0.3513, + "step": 20260 + }, + { + "epoch": 0.3161851874961003, + "grad_norm": 1.4705764055252075, + "learning_rate": 0.00017892202741207546, + "loss": 0.2162, + "step": 20270 + }, + { + "epoch": 0.31634117426842207, + "grad_norm": 2.5773000717163086, + "learning_rate": 0.00017891162829392068, + "loss": 0.1441, + "step": 20280 + }, + { + "epoch": 0.31649716104074377, + "grad_norm": 0.6309418082237244, + "learning_rate": 0.0001789012291757659, + "loss": 0.1477, + "step": 20290 + }, + { + "epoch": 0.31665314781306547, + "grad_norm": 4.573570728302002, + "learning_rate": 0.00017889083005761112, + "loss": 0.3089, + "step": 20300 + }, + { + "epoch": 0.3168091345853872, + "grad_norm": 0.8413357138633728, + "learning_rate": 0.00017888043093945634, + "loss": 0.2751, + "step": 20310 + }, + { + "epoch": 0.3169651213577089, + "grad_norm": 0.9415755271911621, + "learning_rate": 0.00017887003182130156, + "loss": 0.2306, + "step": 20320 + }, + { + "epoch": 0.3171211081300306, + "grad_norm": 0.8843361139297485, + "learning_rate": 0.00017885963270314678, + "loss": 0.1959, + "step": 20330 + }, + { + "epoch": 0.3172770949023523, + "grad_norm": 0.4617684483528137, + "learning_rate": 0.000178849233584992, + "loss": 0.1631, + "step": 20340 + }, + { + "epoch": 0.317433081674674, + "grad_norm": 3.2093448638916016, + "learning_rate": 0.00017883883446683722, + "loss": 0.3872, + "step": 20350 + }, + { + "epoch": 0.3175890684469957, + "grad_norm": 2.095564603805542, + "learning_rate": 0.00017882843534868244, + "loss": 0.2312, + "step": 20360 + }, + { + "epoch": 0.3177450552193174, + "grad_norm": 0.7927945852279663, + "learning_rate": 0.00017881803623052766, + "loss": 0.2254, + "step": 20370 + }, + { + "epoch": 0.3179010419916391, + "grad_norm": 1.9639893770217896, + "learning_rate": 0.00017880763711237288, + "loss": 0.2782, + "step": 20380 + }, + { + "epoch": 0.3180570287639608, + "grad_norm": 0.5046456456184387, + "learning_rate": 0.0001787972379942181, + "loss": 0.2609, + "step": 20390 + }, + { + "epoch": 0.3182130155362825, + "grad_norm": 3.1301658153533936, + "learning_rate": 0.00017878683887606332, + "loss": 0.2682, + "step": 20400 + }, + { + "epoch": 0.3183690023086042, + "grad_norm": 4.775171756744385, + "learning_rate": 0.00017877643975790854, + "loss": 0.4239, + "step": 20410 + }, + { + "epoch": 0.3185249890809259, + "grad_norm": 0.7188621759414673, + "learning_rate": 0.00017876604063975376, + "loss": 0.1219, + "step": 20420 + }, + { + "epoch": 0.31868097585324767, + "grad_norm": 2.135305404663086, + "learning_rate": 0.00017875564152159898, + "loss": 0.1525, + "step": 20430 + }, + { + "epoch": 0.3188369626255694, + "grad_norm": 5.73899507522583, + "learning_rate": 0.0001787452424034442, + "loss": 0.3293, + "step": 20440 + }, + { + "epoch": 0.3189929493978911, + "grad_norm": 0.7671763896942139, + "learning_rate": 0.00017873484328528942, + "loss": 0.3815, + "step": 20450 + }, + { + "epoch": 0.3191489361702128, + "grad_norm": 0.38267794251441956, + "learning_rate": 0.00017872444416713464, + "loss": 0.2484, + "step": 20460 + }, + { + "epoch": 0.3193049229425345, + "grad_norm": 1.1240700483322144, + "learning_rate": 0.00017871404504897986, + "loss": 0.4274, + "step": 20470 + }, + { + "epoch": 0.3194609097148562, + "grad_norm": 1.7546634674072266, + "learning_rate": 0.00017870364593082508, + "loss": 0.2651, + "step": 20480 + }, + { + "epoch": 0.3196168964871779, + "grad_norm": 0.6864067912101746, + "learning_rate": 0.0001786932468126703, + "loss": 0.1648, + "step": 20490 + }, + { + "epoch": 0.3197728832594996, + "grad_norm": 1.5857524871826172, + "learning_rate": 0.00017868284769451552, + "loss": 0.1917, + "step": 20500 + }, + { + "epoch": 0.3199288700318213, + "grad_norm": 1.8687770366668701, + "learning_rate": 0.00017867244857636073, + "loss": 0.3454, + "step": 20510 + }, + { + "epoch": 0.320084856804143, + "grad_norm": 2.078031301498413, + "learning_rate": 0.00017866204945820595, + "loss": 0.232, + "step": 20520 + }, + { + "epoch": 0.3202408435764647, + "grad_norm": 1.000463604927063, + "learning_rate": 0.00017865165034005117, + "loss": 0.1717, + "step": 20530 + }, + { + "epoch": 0.3203968303487864, + "grad_norm": 2.772836685180664, + "learning_rate": 0.0001786412512218964, + "loss": 0.4126, + "step": 20540 + }, + { + "epoch": 0.3205528171211081, + "grad_norm": 1.9367743730545044, + "learning_rate": 0.0001786308521037416, + "loss": 0.3167, + "step": 20550 + }, + { + "epoch": 0.3207088038934298, + "grad_norm": 0.07297099381685257, + "learning_rate": 0.00017862045298558683, + "loss": 0.458, + "step": 20560 + }, + { + "epoch": 0.3208647906657515, + "grad_norm": 1.7589532136917114, + "learning_rate": 0.00017861005386743205, + "loss": 0.094, + "step": 20570 + }, + { + "epoch": 0.3210207774380732, + "grad_norm": 1.2285284996032715, + "learning_rate": 0.00017859965474927727, + "loss": 0.3907, + "step": 20580 + }, + { + "epoch": 0.321176764210395, + "grad_norm": 4.366521835327148, + "learning_rate": 0.0001785892556311225, + "loss": 0.2066, + "step": 20590 + }, + { + "epoch": 0.3213327509827167, + "grad_norm": 3.798428773880005, + "learning_rate": 0.0001785788565129677, + "loss": 0.4166, + "step": 20600 + }, + { + "epoch": 0.3214887377550384, + "grad_norm": 1.5016762018203735, + "learning_rate": 0.00017856845739481293, + "loss": 0.3327, + "step": 20610 + }, + { + "epoch": 0.3216447245273601, + "grad_norm": 0.4517001211643219, + "learning_rate": 0.00017855805827665815, + "loss": 0.2359, + "step": 20620 + }, + { + "epoch": 0.3218007112996818, + "grad_norm": 2.6482317447662354, + "learning_rate": 0.00017854765915850337, + "loss": 0.3978, + "step": 20630 + }, + { + "epoch": 0.3219566980720035, + "grad_norm": 1.3083317279815674, + "learning_rate": 0.0001785372600403486, + "loss": 0.3161, + "step": 20640 + }, + { + "epoch": 0.3221126848443252, + "grad_norm": 1.145427942276001, + "learning_rate": 0.0001785268609221938, + "loss": 0.2105, + "step": 20650 + }, + { + "epoch": 0.3222686716166469, + "grad_norm": 0.4556472599506378, + "learning_rate": 0.00017851646180403903, + "loss": 0.4084, + "step": 20660 + }, + { + "epoch": 0.3224246583889686, + "grad_norm": 2.27005672454834, + "learning_rate": 0.00017850606268588425, + "loss": 0.1423, + "step": 20670 + }, + { + "epoch": 0.3225806451612903, + "grad_norm": 3.4586832523345947, + "learning_rate": 0.00017849566356772947, + "loss": 0.4497, + "step": 20680 + }, + { + "epoch": 0.322736631933612, + "grad_norm": 1.3652698993682861, + "learning_rate": 0.00017848526444957466, + "loss": 0.2917, + "step": 20690 + }, + { + "epoch": 0.3228926187059337, + "grad_norm": 6.490371227264404, + "learning_rate": 0.0001784748653314199, + "loss": 0.3728, + "step": 20700 + }, + { + "epoch": 0.3230486054782554, + "grad_norm": 11.09431266784668, + "learning_rate": 0.0001784644662132651, + "loss": 0.2322, + "step": 20710 + }, + { + "epoch": 0.3232045922505771, + "grad_norm": 0.3489278554916382, + "learning_rate": 0.00017845406709511035, + "loss": 0.1326, + "step": 20720 + }, + { + "epoch": 0.32336057902289883, + "grad_norm": 0.6457053422927856, + "learning_rate": 0.00017844366797695554, + "loss": 0.1412, + "step": 20730 + }, + { + "epoch": 0.3235165657952206, + "grad_norm": 1.2095884084701538, + "learning_rate": 0.0001784332688588008, + "loss": 0.373, + "step": 20740 + }, + { + "epoch": 0.3236725525675423, + "grad_norm": 1.0281665325164795, + "learning_rate": 0.00017842286974064598, + "loss": 0.2973, + "step": 20750 + }, + { + "epoch": 0.323828539339864, + "grad_norm": 1.4616873264312744, + "learning_rate": 0.00017841247062249123, + "loss": 0.4228, + "step": 20760 + }, + { + "epoch": 0.3239845261121857, + "grad_norm": 1.5599627494812012, + "learning_rate": 0.00017840207150433645, + "loss": 0.2367, + "step": 20770 + }, + { + "epoch": 0.3241405128845074, + "grad_norm": 0.5858798027038574, + "learning_rate": 0.00017839167238618167, + "loss": 0.2332, + "step": 20780 + }, + { + "epoch": 0.3242964996568291, + "grad_norm": 1.9172471761703491, + "learning_rate": 0.00017838127326802688, + "loss": 0.3631, + "step": 20790 + }, + { + "epoch": 0.3244524864291508, + "grad_norm": 1.5015684366226196, + "learning_rate": 0.0001783708741498721, + "loss": 0.113, + "step": 20800 + }, + { + "epoch": 0.3246084732014725, + "grad_norm": 1.4167301654815674, + "learning_rate": 0.00017836047503171732, + "loss": 0.1659, + "step": 20810 + }, + { + "epoch": 0.3247644599737942, + "grad_norm": 6.392453670501709, + "learning_rate": 0.00017835007591356254, + "loss": 0.217, + "step": 20820 + }, + { + "epoch": 0.3249204467461159, + "grad_norm": 0.22701123356819153, + "learning_rate": 0.00017833967679540776, + "loss": 0.2353, + "step": 20830 + }, + { + "epoch": 0.3250764335184376, + "grad_norm": 2.8553898334503174, + "learning_rate": 0.00017832927767725298, + "loss": 0.4051, + "step": 20840 + }, + { + "epoch": 0.3252324202907593, + "grad_norm": 2.4016971588134766, + "learning_rate": 0.0001783188785590982, + "loss": 0.3483, + "step": 20850 + }, + { + "epoch": 0.32538840706308103, + "grad_norm": 1.9661591053009033, + "learning_rate": 0.00017830847944094342, + "loss": 0.2153, + "step": 20860 + }, + { + "epoch": 0.32554439383540273, + "grad_norm": 3.4744138717651367, + "learning_rate": 0.00017829808032278864, + "loss": 0.3371, + "step": 20870 + }, + { + "epoch": 0.32570038060772444, + "grad_norm": 0.9500216245651245, + "learning_rate": 0.00017828768120463386, + "loss": 0.2843, + "step": 20880 + }, + { + "epoch": 0.3258563673800462, + "grad_norm": 1.0460073947906494, + "learning_rate": 0.00017827728208647908, + "loss": 0.2056, + "step": 20890 + }, + { + "epoch": 0.3260123541523679, + "grad_norm": 2.223250389099121, + "learning_rate": 0.0001782668829683243, + "loss": 0.2585, + "step": 20900 + }, + { + "epoch": 0.3261683409246896, + "grad_norm": 4.212105751037598, + "learning_rate": 0.00017825648385016952, + "loss": 0.3496, + "step": 20910 + }, + { + "epoch": 0.3263243276970113, + "grad_norm": 0.6845341920852661, + "learning_rate": 0.00017824608473201474, + "loss": 0.2426, + "step": 20920 + }, + { + "epoch": 0.326480314469333, + "grad_norm": 1.270037055015564, + "learning_rate": 0.00017823568561385996, + "loss": 0.1817, + "step": 20930 + }, + { + "epoch": 0.3266363012416547, + "grad_norm": 0.7929283976554871, + "learning_rate": 0.00017822528649570518, + "loss": 0.1664, + "step": 20940 + }, + { + "epoch": 0.3267922880139764, + "grad_norm": 3.9440627098083496, + "learning_rate": 0.0001782148873775504, + "loss": 0.214, + "step": 20950 + }, + { + "epoch": 0.3269482747862981, + "grad_norm": 0.4081120491027832, + "learning_rate": 0.00017820448825939562, + "loss": 0.4212, + "step": 20960 + }, + { + "epoch": 0.3271042615586198, + "grad_norm": 0.5594578981399536, + "learning_rate": 0.00017819408914124084, + "loss": 0.2965, + "step": 20970 + }, + { + "epoch": 0.32726024833094153, + "grad_norm": 2.206573009490967, + "learning_rate": 0.00017818369002308606, + "loss": 0.4426, + "step": 20980 + }, + { + "epoch": 0.32741623510326323, + "grad_norm": 3.0777409076690674, + "learning_rate": 0.00017817329090493128, + "loss": 0.4783, + "step": 20990 + }, + { + "epoch": 0.32757222187558493, + "grad_norm": 5.460344314575195, + "learning_rate": 0.0001781628917867765, + "loss": 0.2472, + "step": 21000 + }, + { + "epoch": 0.32772820864790664, + "grad_norm": 2.182574510574341, + "learning_rate": 0.00017815249266862172, + "loss": 0.3167, + "step": 21010 + }, + { + "epoch": 0.32788419542022834, + "grad_norm": 2.6626338958740234, + "learning_rate": 0.00017814209355046694, + "loss": 0.2616, + "step": 21020 + }, + { + "epoch": 0.32804018219255004, + "grad_norm": 1.088391661643982, + "learning_rate": 0.00017813169443231216, + "loss": 0.5037, + "step": 21030 + }, + { + "epoch": 0.3281961689648718, + "grad_norm": 2.316967725753784, + "learning_rate": 0.00017812129531415738, + "loss": 0.2276, + "step": 21040 + }, + { + "epoch": 0.3283521557371935, + "grad_norm": 1.1988463401794434, + "learning_rate": 0.0001781108961960026, + "loss": 0.257, + "step": 21050 + }, + { + "epoch": 0.3285081425095152, + "grad_norm": 0.4767325520515442, + "learning_rate": 0.00017810049707784782, + "loss": 0.2694, + "step": 21060 + }, + { + "epoch": 0.3286641292818369, + "grad_norm": 0.7175318598747253, + "learning_rate": 0.00017809009795969303, + "loss": 0.289, + "step": 21070 + }, + { + "epoch": 0.3288201160541586, + "grad_norm": 1.733170747756958, + "learning_rate": 0.00017807969884153825, + "loss": 0.4186, + "step": 21080 + }, + { + "epoch": 0.3289761028264803, + "grad_norm": 1.299400806427002, + "learning_rate": 0.00017806929972338347, + "loss": 0.2, + "step": 21090 + }, + { + "epoch": 0.329132089598802, + "grad_norm": 1.8722649812698364, + "learning_rate": 0.0001780589006052287, + "loss": 0.3041, + "step": 21100 + }, + { + "epoch": 0.32928807637112373, + "grad_norm": 2.6847896575927734, + "learning_rate": 0.0001780485014870739, + "loss": 0.2804, + "step": 21110 + }, + { + "epoch": 0.32944406314344543, + "grad_norm": 2.485938310623169, + "learning_rate": 0.00017803810236891913, + "loss": 0.2618, + "step": 21120 + }, + { + "epoch": 0.32960004991576713, + "grad_norm": 1.780991554260254, + "learning_rate": 0.00017802770325076435, + "loss": 0.2841, + "step": 21130 + }, + { + "epoch": 0.32975603668808884, + "grad_norm": 1.3728632926940918, + "learning_rate": 0.00017801730413260955, + "loss": 0.1419, + "step": 21140 + }, + { + "epoch": 0.32991202346041054, + "grad_norm": 1.587583303451538, + "learning_rate": 0.0001780069050144548, + "loss": 0.2173, + "step": 21150 + }, + { + "epoch": 0.33006801023273225, + "grad_norm": 1.8564294576644897, + "learning_rate": 0.00017799650589629998, + "loss": 0.3012, + "step": 21160 + }, + { + "epoch": 0.33022399700505395, + "grad_norm": 0.22945725917816162, + "learning_rate": 0.00017798610677814523, + "loss": 0.2146, + "step": 21170 + }, + { + "epoch": 0.33037998377737565, + "grad_norm": 0.5251349210739136, + "learning_rate": 0.00017797570765999042, + "loss": 0.3422, + "step": 21180 + }, + { + "epoch": 0.3305359705496974, + "grad_norm": 1.1029380559921265, + "learning_rate": 0.00017796530854183567, + "loss": 0.3739, + "step": 21190 + }, + { + "epoch": 0.3306919573220191, + "grad_norm": 0.18028827011585236, + "learning_rate": 0.00017795490942368086, + "loss": 0.3563, + "step": 21200 + }, + { + "epoch": 0.3308479440943408, + "grad_norm": 3.026299476623535, + "learning_rate": 0.0001779445103055261, + "loss": 0.1437, + "step": 21210 + }, + { + "epoch": 0.3310039308666625, + "grad_norm": 0.5276438593864441, + "learning_rate": 0.0001779341111873713, + "loss": 0.3588, + "step": 21220 + }, + { + "epoch": 0.3311599176389842, + "grad_norm": 0.838901698589325, + "learning_rate": 0.00017792371206921655, + "loss": 0.3944, + "step": 21230 + }, + { + "epoch": 0.33131590441130593, + "grad_norm": 0.7526358366012573, + "learning_rate": 0.00017791331295106174, + "loss": 0.3618, + "step": 21240 + }, + { + "epoch": 0.33147189118362763, + "grad_norm": 0.34405943751335144, + "learning_rate": 0.000177902913832907, + "loss": 0.168, + "step": 21250 + }, + { + "epoch": 0.33162787795594934, + "grad_norm": 1.2651864290237427, + "learning_rate": 0.00017789251471475218, + "loss": 0.2877, + "step": 21260 + }, + { + "epoch": 0.33178386472827104, + "grad_norm": 0.4226759374141693, + "learning_rate": 0.00017788211559659743, + "loss": 0.2135, + "step": 21270 + }, + { + "epoch": 0.33193985150059274, + "grad_norm": 1.4860749244689941, + "learning_rate": 0.00017787171647844262, + "loss": 0.4945, + "step": 21280 + }, + { + "epoch": 0.33209583827291445, + "grad_norm": 3.712172746658325, + "learning_rate": 0.00017786131736028787, + "loss": 0.4228, + "step": 21290 + }, + { + "epoch": 0.33225182504523615, + "grad_norm": 0.12535716593265533, + "learning_rate": 0.00017785091824213306, + "loss": 0.1053, + "step": 21300 + }, + { + "epoch": 0.33240781181755785, + "grad_norm": 1.6500813961029053, + "learning_rate": 0.0001778405191239783, + "loss": 0.3572, + "step": 21310 + }, + { + "epoch": 0.33256379858987956, + "grad_norm": 2.3633081912994385, + "learning_rate": 0.0001778301200058235, + "loss": 0.3679, + "step": 21320 + }, + { + "epoch": 0.33271978536220126, + "grad_norm": 0.5572711229324341, + "learning_rate": 0.00017781972088766875, + "loss": 0.2661, + "step": 21330 + }, + { + "epoch": 0.332875772134523, + "grad_norm": 1.1650604009628296, + "learning_rate": 0.00017780932176951394, + "loss": 0.2094, + "step": 21340 + }, + { + "epoch": 0.3330317589068447, + "grad_norm": 0.7077860236167908, + "learning_rate": 0.00017779892265135918, + "loss": 0.149, + "step": 21350 + }, + { + "epoch": 0.3331877456791664, + "grad_norm": 0.4741782546043396, + "learning_rate": 0.00017778852353320438, + "loss": 0.1405, + "step": 21360 + }, + { + "epoch": 0.33334373245148813, + "grad_norm": 5.10825777053833, + "learning_rate": 0.00017777812441504962, + "loss": 0.2863, + "step": 21370 + }, + { + "epoch": 0.33349971922380983, + "grad_norm": 0.2401765137910843, + "learning_rate": 0.00017776772529689482, + "loss": 0.3138, + "step": 21380 + }, + { + "epoch": 0.33365570599613154, + "grad_norm": 2.0802645683288574, + "learning_rate": 0.00017775732617874006, + "loss": 0.2106, + "step": 21390 + }, + { + "epoch": 0.33381169276845324, + "grad_norm": 6.736307621002197, + "learning_rate": 0.00017774692706058526, + "loss": 0.1115, + "step": 21400 + }, + { + "epoch": 0.33396767954077494, + "grad_norm": 1.9653793573379517, + "learning_rate": 0.0001777365279424305, + "loss": 0.3681, + "step": 21410 + }, + { + "epoch": 0.33412366631309665, + "grad_norm": 0.260010302066803, + "learning_rate": 0.0001777261288242757, + "loss": 0.2099, + "step": 21420 + }, + { + "epoch": 0.33427965308541835, + "grad_norm": 1.4657262563705444, + "learning_rate": 0.00017771572970612094, + "loss": 0.2862, + "step": 21430 + }, + { + "epoch": 0.33443563985774005, + "grad_norm": 0.2749074101448059, + "learning_rate": 0.00017770533058796613, + "loss": 0.2737, + "step": 21440 + }, + { + "epoch": 0.33459162663006176, + "grad_norm": 0.8197032809257507, + "learning_rate": 0.00017769493146981138, + "loss": 0.1465, + "step": 21450 + }, + { + "epoch": 0.33474761340238346, + "grad_norm": 0.2702539563179016, + "learning_rate": 0.00017768453235165657, + "loss": 0.1653, + "step": 21460 + }, + { + "epoch": 0.33490360017470516, + "grad_norm": 2.730003595352173, + "learning_rate": 0.00017767413323350182, + "loss": 0.4171, + "step": 21470 + }, + { + "epoch": 0.33505958694702687, + "grad_norm": 1.7378095388412476, + "learning_rate": 0.000177663734115347, + "loss": 0.2022, + "step": 21480 + }, + { + "epoch": 0.3352155737193486, + "grad_norm": 3.3789122104644775, + "learning_rate": 0.00017765333499719226, + "loss": 0.3737, + "step": 21490 + }, + { + "epoch": 0.33537156049167033, + "grad_norm": 0.83452969789505, + "learning_rate": 0.00017764293587903745, + "loss": 0.2577, + "step": 21500 + }, + { + "epoch": 0.33552754726399203, + "grad_norm": 1.0536013841629028, + "learning_rate": 0.0001776325367608827, + "loss": 0.4947, + "step": 21510 + }, + { + "epoch": 0.33568353403631374, + "grad_norm": 0.0694153904914856, + "learning_rate": 0.0001776221376427279, + "loss": 0.1109, + "step": 21520 + }, + { + "epoch": 0.33583952080863544, + "grad_norm": 1.87344229221344, + "learning_rate": 0.00017761173852457314, + "loss": 0.2946, + "step": 21530 + }, + { + "epoch": 0.33599550758095714, + "grad_norm": 1.3620704412460327, + "learning_rate": 0.00017760133940641833, + "loss": 0.3758, + "step": 21540 + }, + { + "epoch": 0.33615149435327885, + "grad_norm": 0.24896468222141266, + "learning_rate": 0.00017759094028826358, + "loss": 0.3, + "step": 21550 + }, + { + "epoch": 0.33630748112560055, + "grad_norm": 1.418351411819458, + "learning_rate": 0.00017758054117010877, + "loss": 0.4662, + "step": 21560 + }, + { + "epoch": 0.33646346789792225, + "grad_norm": 0.8420088887214661, + "learning_rate": 0.00017757014205195402, + "loss": 0.1271, + "step": 21570 + }, + { + "epoch": 0.33661945467024396, + "grad_norm": 2.7175440788269043, + "learning_rate": 0.0001775597429337992, + "loss": 0.3309, + "step": 21580 + }, + { + "epoch": 0.33677544144256566, + "grad_norm": 1.5857981443405151, + "learning_rate": 0.00017754934381564443, + "loss": 0.2379, + "step": 21590 + }, + { + "epoch": 0.33693142821488736, + "grad_norm": 0.6439513564109802, + "learning_rate": 0.00017753894469748965, + "loss": 0.2321, + "step": 21600 + }, + { + "epoch": 0.33708741498720907, + "grad_norm": 2.5145156383514404, + "learning_rate": 0.00017752854557933487, + "loss": 0.2183, + "step": 21610 + }, + { + "epoch": 0.33724340175953077, + "grad_norm": 0.5154660940170288, + "learning_rate": 0.0001775181464611801, + "loss": 0.3879, + "step": 21620 + }, + { + "epoch": 0.3373993885318525, + "grad_norm": 2.2567896842956543, + "learning_rate": 0.0001775077473430253, + "loss": 0.2931, + "step": 21630 + }, + { + "epoch": 0.33755537530417423, + "grad_norm": 1.013460636138916, + "learning_rate": 0.00017749734822487055, + "loss": 0.2663, + "step": 21640 + }, + { + "epoch": 0.33771136207649594, + "grad_norm": 4.204822063446045, + "learning_rate": 0.00017748694910671575, + "loss": 0.4413, + "step": 21650 + }, + { + "epoch": 0.33786734884881764, + "grad_norm": 3.314096212387085, + "learning_rate": 0.000177476549988561, + "loss": 0.2016, + "step": 21660 + }, + { + "epoch": 0.33802333562113934, + "grad_norm": 7.210069179534912, + "learning_rate": 0.00017746615087040619, + "loss": 0.3756, + "step": 21670 + }, + { + "epoch": 0.33817932239346105, + "grad_norm": 0.5357813835144043, + "learning_rate": 0.00017745575175225143, + "loss": 0.0972, + "step": 21680 + }, + { + "epoch": 0.33833530916578275, + "grad_norm": 1.383971095085144, + "learning_rate": 0.00017744535263409663, + "loss": 0.1316, + "step": 21690 + }, + { + "epoch": 0.33849129593810445, + "grad_norm": 0.829406201839447, + "learning_rate": 0.00017743495351594187, + "loss": 0.1488, + "step": 21700 + }, + { + "epoch": 0.33864728271042616, + "grad_norm": 1.8377529382705688, + "learning_rate": 0.00017742455439778706, + "loss": 0.3737, + "step": 21710 + }, + { + "epoch": 0.33880326948274786, + "grad_norm": 0.026477506384253502, + "learning_rate": 0.0001774141552796323, + "loss": 0.1839, + "step": 21720 + }, + { + "epoch": 0.33895925625506956, + "grad_norm": 1.9650471210479736, + "learning_rate": 0.0001774037561614775, + "loss": 0.4044, + "step": 21730 + }, + { + "epoch": 0.33911524302739127, + "grad_norm": 0.3597981929779053, + "learning_rate": 0.00017739335704332275, + "loss": 0.2774, + "step": 21740 + }, + { + "epoch": 0.33927122979971297, + "grad_norm": 2.270568370819092, + "learning_rate": 0.00017738295792516794, + "loss": 0.1042, + "step": 21750 + }, + { + "epoch": 0.3394272165720347, + "grad_norm": 2.03104567527771, + "learning_rate": 0.0001773725588070132, + "loss": 0.1452, + "step": 21760 + }, + { + "epoch": 0.3395832033443564, + "grad_norm": 2.912268877029419, + "learning_rate": 0.00017736215968885838, + "loss": 0.4327, + "step": 21770 + }, + { + "epoch": 0.3397391901166781, + "grad_norm": 2.9658117294311523, + "learning_rate": 0.00017735176057070363, + "loss": 0.197, + "step": 21780 + }, + { + "epoch": 0.33989517688899984, + "grad_norm": 5.578575611114502, + "learning_rate": 0.00017734136145254882, + "loss": 0.21, + "step": 21790 + }, + { + "epoch": 0.34005116366132154, + "grad_norm": 1.8583189249038696, + "learning_rate": 0.00017733096233439407, + "loss": 0.3938, + "step": 21800 + }, + { + "epoch": 0.34020715043364325, + "grad_norm": 6.439654350280762, + "learning_rate": 0.00017732056321623926, + "loss": 0.3358, + "step": 21810 + }, + { + "epoch": 0.34036313720596495, + "grad_norm": 0.4864749014377594, + "learning_rate": 0.0001773101640980845, + "loss": 0.2587, + "step": 21820 + }, + { + "epoch": 0.34051912397828665, + "grad_norm": 0.20829786360263824, + "learning_rate": 0.0001772997649799297, + "loss": 0.1931, + "step": 21830 + }, + { + "epoch": 0.34067511075060836, + "grad_norm": 2.3870739936828613, + "learning_rate": 0.00017728936586177495, + "loss": 0.2562, + "step": 21840 + }, + { + "epoch": 0.34083109752293006, + "grad_norm": 6.137410640716553, + "learning_rate": 0.00017727896674362014, + "loss": 0.2905, + "step": 21850 + }, + { + "epoch": 0.34098708429525176, + "grad_norm": 0.2211449146270752, + "learning_rate": 0.0001772685676254654, + "loss": 0.1578, + "step": 21860 + }, + { + "epoch": 0.34114307106757347, + "grad_norm": 2.6646316051483154, + "learning_rate": 0.00017725816850731058, + "loss": 0.1941, + "step": 21870 + }, + { + "epoch": 0.34129905783989517, + "grad_norm": 0.651335597038269, + "learning_rate": 0.00017724776938915583, + "loss": 0.1766, + "step": 21880 + }, + { + "epoch": 0.3414550446122169, + "grad_norm": 3.878089427947998, + "learning_rate": 0.00017723737027100102, + "loss": 0.2258, + "step": 21890 + }, + { + "epoch": 0.3416110313845386, + "grad_norm": 2.8596532344818115, + "learning_rate": 0.00017722697115284627, + "loss": 0.224, + "step": 21900 + }, + { + "epoch": 0.3417670181568603, + "grad_norm": 0.16343307495117188, + "learning_rate": 0.00017721657203469146, + "loss": 0.1466, + "step": 21910 + }, + { + "epoch": 0.341923004929182, + "grad_norm": 1.0998289585113525, + "learning_rate": 0.0001772061729165367, + "loss": 0.1408, + "step": 21920 + }, + { + "epoch": 0.3420789917015037, + "grad_norm": 0.9425449967384338, + "learning_rate": 0.0001771957737983819, + "loss": 0.2038, + "step": 21930 + }, + { + "epoch": 0.34223497847382545, + "grad_norm": 1.8072413206100464, + "learning_rate": 0.00017718537468022714, + "loss": 0.4461, + "step": 21940 + }, + { + "epoch": 0.34239096524614715, + "grad_norm": 0.7303923964500427, + "learning_rate": 0.00017717497556207234, + "loss": 0.1389, + "step": 21950 + }, + { + "epoch": 0.34254695201846885, + "grad_norm": 1.2838555574417114, + "learning_rate": 0.00017716457644391758, + "loss": 0.4434, + "step": 21960 + }, + { + "epoch": 0.34270293879079056, + "grad_norm": 0.3216365873813629, + "learning_rate": 0.00017715417732576278, + "loss": 0.2329, + "step": 21970 + }, + { + "epoch": 0.34285892556311226, + "grad_norm": 0.30308425426483154, + "learning_rate": 0.00017714377820760802, + "loss": 0.2014, + "step": 21980 + }, + { + "epoch": 0.34301491233543396, + "grad_norm": 2.267559051513672, + "learning_rate": 0.00017713337908945321, + "loss": 0.0936, + "step": 21990 + }, + { + "epoch": 0.34317089910775567, + "grad_norm": 1.0513626337051392, + "learning_rate": 0.00017712297997129846, + "loss": 0.2288, + "step": 22000 + }, + { + "epoch": 0.34332688588007737, + "grad_norm": 0.8992903828620911, + "learning_rate": 0.00017711258085314365, + "loss": 0.3064, + "step": 22010 + }, + { + "epoch": 0.3434828726523991, + "grad_norm": 1.0323415994644165, + "learning_rate": 0.0001771021817349889, + "loss": 0.2251, + "step": 22020 + }, + { + "epoch": 0.3436388594247208, + "grad_norm": 1.361558198928833, + "learning_rate": 0.0001770917826168341, + "loss": 0.1273, + "step": 22030 + }, + { + "epoch": 0.3437948461970425, + "grad_norm": 0.6209410429000854, + "learning_rate": 0.0001770813834986793, + "loss": 0.1759, + "step": 22040 + }, + { + "epoch": 0.3439508329693642, + "grad_norm": 0.09486691653728485, + "learning_rate": 0.00017707098438052453, + "loss": 0.2522, + "step": 22050 + }, + { + "epoch": 0.3441068197416859, + "grad_norm": 2.731093645095825, + "learning_rate": 0.00017706058526236975, + "loss": 0.4172, + "step": 22060 + }, + { + "epoch": 0.3442628065140076, + "grad_norm": 1.4385517835617065, + "learning_rate": 0.00017705018614421497, + "loss": 0.1646, + "step": 22070 + }, + { + "epoch": 0.3444187932863293, + "grad_norm": 1.858129858970642, + "learning_rate": 0.0001770397870260602, + "loss": 0.1447, + "step": 22080 + }, + { + "epoch": 0.34457478005865105, + "grad_norm": 0.2235264778137207, + "learning_rate": 0.0001770293879079054, + "loss": 0.1518, + "step": 22090 + }, + { + "epoch": 0.34473076683097276, + "grad_norm": 1.930906057357788, + "learning_rate": 0.00017701898878975063, + "loss": 0.2362, + "step": 22100 + }, + { + "epoch": 0.34488675360329446, + "grad_norm": 0.8461508750915527, + "learning_rate": 0.00017700858967159585, + "loss": 0.1288, + "step": 22110 + }, + { + "epoch": 0.34504274037561616, + "grad_norm": 0.09053146094083786, + "learning_rate": 0.00017699819055344107, + "loss": 0.306, + "step": 22120 + }, + { + "epoch": 0.34519872714793787, + "grad_norm": 1.3184669017791748, + "learning_rate": 0.0001769877914352863, + "loss": 0.1227, + "step": 22130 + }, + { + "epoch": 0.34535471392025957, + "grad_norm": 1.534020185470581, + "learning_rate": 0.0001769773923171315, + "loss": 0.1278, + "step": 22140 + }, + { + "epoch": 0.3455107006925813, + "grad_norm": 3.5368030071258545, + "learning_rate": 0.00017696699319897673, + "loss": 0.5032, + "step": 22150 + }, + { + "epoch": 0.345666687464903, + "grad_norm": 0.8782071471214294, + "learning_rate": 0.00017695659408082195, + "loss": 0.2905, + "step": 22160 + }, + { + "epoch": 0.3458226742372247, + "grad_norm": 0.30514800548553467, + "learning_rate": 0.00017694619496266717, + "loss": 0.205, + "step": 22170 + }, + { + "epoch": 0.3459786610095464, + "grad_norm": 1.0300681591033936, + "learning_rate": 0.0001769357958445124, + "loss": 0.2816, + "step": 22180 + }, + { + "epoch": 0.3461346477818681, + "grad_norm": 0.4893283247947693, + "learning_rate": 0.0001769253967263576, + "loss": 0.2256, + "step": 22190 + }, + { + "epoch": 0.3462906345541898, + "grad_norm": 2.1289849281311035, + "learning_rate": 0.00017691499760820283, + "loss": 0.1921, + "step": 22200 + }, + { + "epoch": 0.3464466213265115, + "grad_norm": 0.11431695520877838, + "learning_rate": 0.00017690459849004805, + "loss": 0.2066, + "step": 22210 + }, + { + "epoch": 0.3466026080988332, + "grad_norm": 3.165898561477661, + "learning_rate": 0.00017689419937189327, + "loss": 0.4442, + "step": 22220 + }, + { + "epoch": 0.3467585948711549, + "grad_norm": 0.6485002040863037, + "learning_rate": 0.00017688380025373849, + "loss": 0.4352, + "step": 22230 + }, + { + "epoch": 0.34691458164347666, + "grad_norm": 0.5462929606437683, + "learning_rate": 0.0001768734011355837, + "loss": 0.1493, + "step": 22240 + }, + { + "epoch": 0.34707056841579836, + "grad_norm": 2.157317876815796, + "learning_rate": 0.00017686300201742893, + "loss": 0.2205, + "step": 22250 + }, + { + "epoch": 0.34722655518812007, + "grad_norm": 2.8473408222198486, + "learning_rate": 0.00017685260289927414, + "loss": 0.3636, + "step": 22260 + }, + { + "epoch": 0.34738254196044177, + "grad_norm": 4.030854225158691, + "learning_rate": 0.00017684220378111936, + "loss": 0.1713, + "step": 22270 + }, + { + "epoch": 0.3475385287327635, + "grad_norm": 4.853302001953125, + "learning_rate": 0.00017683180466296458, + "loss": 0.3476, + "step": 22280 + }, + { + "epoch": 0.3476945155050852, + "grad_norm": 0.8569431304931641, + "learning_rate": 0.0001768214055448098, + "loss": 0.1835, + "step": 22290 + }, + { + "epoch": 0.3478505022774069, + "grad_norm": 2.213550567626953, + "learning_rate": 0.00017681100642665502, + "loss": 0.1956, + "step": 22300 + }, + { + "epoch": 0.3480064890497286, + "grad_norm": 0.07390352338552475, + "learning_rate": 0.00017680060730850024, + "loss": 0.3881, + "step": 22310 + }, + { + "epoch": 0.3481624758220503, + "grad_norm": 0.5976702570915222, + "learning_rate": 0.00017679020819034546, + "loss": 0.1883, + "step": 22320 + }, + { + "epoch": 0.348318462594372, + "grad_norm": 2.6148743629455566, + "learning_rate": 0.00017677980907219068, + "loss": 0.1151, + "step": 22330 + }, + { + "epoch": 0.3484744493666937, + "grad_norm": 0.9557235240936279, + "learning_rate": 0.0001767694099540359, + "loss": 0.1261, + "step": 22340 + }, + { + "epoch": 0.3486304361390154, + "grad_norm": 3.4990904331207275, + "learning_rate": 0.00017675901083588112, + "loss": 0.397, + "step": 22350 + }, + { + "epoch": 0.3487864229113371, + "grad_norm": 3.0992941856384277, + "learning_rate": 0.00017674861171772634, + "loss": 0.1306, + "step": 22360 + }, + { + "epoch": 0.3489424096836588, + "grad_norm": 0.9098914861679077, + "learning_rate": 0.00017673821259957156, + "loss": 0.265, + "step": 22370 + }, + { + "epoch": 0.3490983964559805, + "grad_norm": 0.6933314800262451, + "learning_rate": 0.00017672781348141678, + "loss": 0.1889, + "step": 22380 + }, + { + "epoch": 0.34925438322830227, + "grad_norm": 1.680850625038147, + "learning_rate": 0.000176717414363262, + "loss": 0.4258, + "step": 22390 + }, + { + "epoch": 0.34941037000062397, + "grad_norm": 0.6795231699943542, + "learning_rate": 0.00017670701524510722, + "loss": 0.2469, + "step": 22400 + }, + { + "epoch": 0.3495663567729457, + "grad_norm": 1.9486218690872192, + "learning_rate": 0.00017669661612695244, + "loss": 0.2466, + "step": 22410 + }, + { + "epoch": 0.3497223435452674, + "grad_norm": 3.4444730281829834, + "learning_rate": 0.00017668621700879766, + "loss": 0.6108, + "step": 22420 + }, + { + "epoch": 0.3498783303175891, + "grad_norm": 1.1965771913528442, + "learning_rate": 0.00017667581789064288, + "loss": 0.1571, + "step": 22430 + }, + { + "epoch": 0.3500343170899108, + "grad_norm": 0.5158458948135376, + "learning_rate": 0.0001766654187724881, + "loss": 0.1704, + "step": 22440 + }, + { + "epoch": 0.3501903038622325, + "grad_norm": 0.7161465883255005, + "learning_rate": 0.00017665501965433332, + "loss": 0.1734, + "step": 22450 + }, + { + "epoch": 0.3503462906345542, + "grad_norm": 0.322613924741745, + "learning_rate": 0.00017664462053617854, + "loss": 0.2174, + "step": 22460 + }, + { + "epoch": 0.3505022774068759, + "grad_norm": 1.2843927145004272, + "learning_rate": 0.00017663422141802376, + "loss": 0.112, + "step": 22470 + }, + { + "epoch": 0.3506582641791976, + "grad_norm": 0.6412551999092102, + "learning_rate": 0.00017662382229986898, + "loss": 0.2097, + "step": 22480 + }, + { + "epoch": 0.3508142509515193, + "grad_norm": 0.4367602467536926, + "learning_rate": 0.0001766134231817142, + "loss": 0.2376, + "step": 22490 + }, + { + "epoch": 0.350970237723841, + "grad_norm": 0.8941527009010315, + "learning_rate": 0.00017660302406355942, + "loss": 0.1876, + "step": 22500 + }, + { + "epoch": 0.3511262244961627, + "grad_norm": 2.340299367904663, + "learning_rate": 0.00017659262494540464, + "loss": 0.1629, + "step": 22510 + }, + { + "epoch": 0.3512822112684844, + "grad_norm": 2.225738286972046, + "learning_rate": 0.00017658222582724986, + "loss": 0.3814, + "step": 22520 + }, + { + "epoch": 0.3514381980408061, + "grad_norm": 1.4277156591415405, + "learning_rate": 0.00017657182670909508, + "loss": 0.3145, + "step": 22530 + }, + { + "epoch": 0.3515941848131278, + "grad_norm": 1.1834521293640137, + "learning_rate": 0.0001765614275909403, + "loss": 0.3575, + "step": 22540 + }, + { + "epoch": 0.3517501715854496, + "grad_norm": 1.9734115600585938, + "learning_rate": 0.00017655102847278551, + "loss": 0.3772, + "step": 22550 + }, + { + "epoch": 0.3519061583577713, + "grad_norm": 2.2094125747680664, + "learning_rate": 0.00017654062935463073, + "loss": 0.3748, + "step": 22560 + }, + { + "epoch": 0.352062145130093, + "grad_norm": 4.1570658683776855, + "learning_rate": 0.00017653023023647595, + "loss": 0.3592, + "step": 22570 + }, + { + "epoch": 0.3522181319024147, + "grad_norm": 0.38426318764686584, + "learning_rate": 0.00017651983111832117, + "loss": 0.238, + "step": 22580 + }, + { + "epoch": 0.3523741186747364, + "grad_norm": 2.443509578704834, + "learning_rate": 0.0001765094320001664, + "loss": 0.2557, + "step": 22590 + }, + { + "epoch": 0.3525301054470581, + "grad_norm": 1.4817020893096924, + "learning_rate": 0.0001764990328820116, + "loss": 0.3088, + "step": 22600 + }, + { + "epoch": 0.3526860922193798, + "grad_norm": 1.3554056882858276, + "learning_rate": 0.00017648863376385683, + "loss": 0.2346, + "step": 22610 + }, + { + "epoch": 0.3528420789917015, + "grad_norm": 0.15621060132980347, + "learning_rate": 0.00017647823464570205, + "loss": 0.2509, + "step": 22620 + }, + { + "epoch": 0.3529980657640232, + "grad_norm": 2.251452684402466, + "learning_rate": 0.00017646783552754727, + "loss": 0.3074, + "step": 22630 + }, + { + "epoch": 0.3531540525363449, + "grad_norm": 0.556117057800293, + "learning_rate": 0.0001764574364093925, + "loss": 0.1189, + "step": 22640 + }, + { + "epoch": 0.3533100393086666, + "grad_norm": 0.8643046617507935, + "learning_rate": 0.0001764470372912377, + "loss": 0.1117, + "step": 22650 + }, + { + "epoch": 0.3534660260809883, + "grad_norm": 2.586470365524292, + "learning_rate": 0.00017643663817308293, + "loss": 0.1658, + "step": 22660 + }, + { + "epoch": 0.35362201285331, + "grad_norm": 0.07640788704156876, + "learning_rate": 0.00017642623905492815, + "loss": 0.2076, + "step": 22670 + }, + { + "epoch": 0.3537779996256317, + "grad_norm": 2.1447646617889404, + "learning_rate": 0.00017641583993677337, + "loss": 0.3952, + "step": 22680 + }, + { + "epoch": 0.3539339863979534, + "grad_norm": 2.0806121826171875, + "learning_rate": 0.0001764054408186186, + "loss": 0.2465, + "step": 22690 + }, + { + "epoch": 0.3540899731702752, + "grad_norm": 0.15225516259670258, + "learning_rate": 0.0001763950417004638, + "loss": 0.2529, + "step": 22700 + }, + { + "epoch": 0.3542459599425969, + "grad_norm": 0.8685635328292847, + "learning_rate": 0.00017638464258230903, + "loss": 0.5026, + "step": 22710 + }, + { + "epoch": 0.3544019467149186, + "grad_norm": 0.43446022272109985, + "learning_rate": 0.00017637424346415425, + "loss": 0.3306, + "step": 22720 + }, + { + "epoch": 0.3545579334872403, + "grad_norm": 0.21245236694812775, + "learning_rate": 0.00017636384434599947, + "loss": 0.1259, + "step": 22730 + }, + { + "epoch": 0.354713920259562, + "grad_norm": 0.5208606123924255, + "learning_rate": 0.0001763534452278447, + "loss": 0.2927, + "step": 22740 + }, + { + "epoch": 0.3548699070318837, + "grad_norm": 0.75662761926651, + "learning_rate": 0.0001763430461096899, + "loss": 0.1677, + "step": 22750 + }, + { + "epoch": 0.3550258938042054, + "grad_norm": 0.7772522568702698, + "learning_rate": 0.00017633264699153513, + "loss": 0.2772, + "step": 22760 + }, + { + "epoch": 0.3551818805765271, + "grad_norm": 1.768527626991272, + "learning_rate": 0.00017632224787338035, + "loss": 0.1803, + "step": 22770 + }, + { + "epoch": 0.3553378673488488, + "grad_norm": 0.5841684937477112, + "learning_rate": 0.00017631184875522557, + "loss": 0.23, + "step": 22780 + }, + { + "epoch": 0.3554938541211705, + "grad_norm": 0.0703231692314148, + "learning_rate": 0.00017630144963707079, + "loss": 0.2504, + "step": 22790 + }, + { + "epoch": 0.3556498408934922, + "grad_norm": 4.701184272766113, + "learning_rate": 0.000176291050518916, + "loss": 0.0678, + "step": 22800 + }, + { + "epoch": 0.3558058276658139, + "grad_norm": 2.3231558799743652, + "learning_rate": 0.00017628065140076123, + "loss": 0.248, + "step": 22810 + }, + { + "epoch": 0.3559618144381356, + "grad_norm": 0.4551299512386322, + "learning_rate": 0.00017627025228260644, + "loss": 0.1933, + "step": 22820 + }, + { + "epoch": 0.35611780121045733, + "grad_norm": 3.0799286365509033, + "learning_rate": 0.00017625985316445166, + "loss": 0.3023, + "step": 22830 + }, + { + "epoch": 0.35627378798277903, + "grad_norm": 1.7619458436965942, + "learning_rate": 0.00017624945404629688, + "loss": 0.3294, + "step": 22840 + }, + { + "epoch": 0.3564297747551008, + "grad_norm": 0.3896751403808594, + "learning_rate": 0.0001762390549281421, + "loss": 0.0388, + "step": 22850 + }, + { + "epoch": 0.3565857615274225, + "grad_norm": 4.727908611297607, + "learning_rate": 0.00017622865580998732, + "loss": 0.2075, + "step": 22860 + }, + { + "epoch": 0.3567417482997442, + "grad_norm": 1.2066168785095215, + "learning_rate": 0.00017621825669183254, + "loss": 0.2297, + "step": 22870 + }, + { + "epoch": 0.3568977350720659, + "grad_norm": 0.4707547724246979, + "learning_rate": 0.00017620785757367776, + "loss": 0.3783, + "step": 22880 + }, + { + "epoch": 0.3570537218443876, + "grad_norm": 1.812171459197998, + "learning_rate": 0.00017619745845552298, + "loss": 0.266, + "step": 22890 + }, + { + "epoch": 0.3572097086167093, + "grad_norm": 1.8720440864562988, + "learning_rate": 0.0001761870593373682, + "loss": 0.152, + "step": 22900 + }, + { + "epoch": 0.357365695389031, + "grad_norm": 0.9550348520278931, + "learning_rate": 0.00017617666021921342, + "loss": 0.3226, + "step": 22910 + }, + { + "epoch": 0.3575216821613527, + "grad_norm": 0.8971384167671204, + "learning_rate": 0.00017616626110105864, + "loss": 0.234, + "step": 22920 + }, + { + "epoch": 0.3576776689336744, + "grad_norm": 2.038457155227661, + "learning_rate": 0.00017615586198290386, + "loss": 0.3064, + "step": 22930 + }, + { + "epoch": 0.3578336557059961, + "grad_norm": 0.052239127457141876, + "learning_rate": 0.00017614546286474905, + "loss": 0.1587, + "step": 22940 + }, + { + "epoch": 0.3579896424783178, + "grad_norm": 0.37484607100486755, + "learning_rate": 0.0001761350637465943, + "loss": 0.248, + "step": 22950 + }, + { + "epoch": 0.35814562925063953, + "grad_norm": 1.459640622138977, + "learning_rate": 0.00017612466462843952, + "loss": 0.3022, + "step": 22960 + }, + { + "epoch": 0.35830161602296123, + "grad_norm": 3.073653221130371, + "learning_rate": 0.00017611426551028474, + "loss": 0.2473, + "step": 22970 + }, + { + "epoch": 0.35845760279528294, + "grad_norm": 1.0391371250152588, + "learning_rate": 0.00017610386639212996, + "loss": 0.1015, + "step": 22980 + }, + { + "epoch": 0.35861358956760464, + "grad_norm": 0.321556955575943, + "learning_rate": 0.00017609346727397518, + "loss": 0.1523, + "step": 22990 + }, + { + "epoch": 0.3587695763399264, + "grad_norm": 6.030311107635498, + "learning_rate": 0.0001760830681558204, + "loss": 0.417, + "step": 23000 + }, + { + "epoch": 0.3589255631122481, + "grad_norm": 0.35989582538604736, + "learning_rate": 0.00017607266903766562, + "loss": 0.2229, + "step": 23010 + }, + { + "epoch": 0.3590815498845698, + "grad_norm": 0.8488921523094177, + "learning_rate": 0.00017606226991951084, + "loss": 0.4153, + "step": 23020 + }, + { + "epoch": 0.3592375366568915, + "grad_norm": 0.8105539083480835, + "learning_rate": 0.00017605187080135606, + "loss": 0.3456, + "step": 23030 + }, + { + "epoch": 0.3593935234292132, + "grad_norm": 0.6721516251564026, + "learning_rate": 0.00017604147168320128, + "loss": 0.4522, + "step": 23040 + }, + { + "epoch": 0.3595495102015349, + "grad_norm": 1.287809133529663, + "learning_rate": 0.0001760310725650465, + "loss": 0.315, + "step": 23050 + }, + { + "epoch": 0.3597054969738566, + "grad_norm": 1.9372718334197998, + "learning_rate": 0.00017602067344689172, + "loss": 0.2472, + "step": 23060 + }, + { + "epoch": 0.3598614837461783, + "grad_norm": 0.0871971994638443, + "learning_rate": 0.00017601027432873694, + "loss": 0.0929, + "step": 23070 + }, + { + "epoch": 0.3600174705185, + "grad_norm": 3.9049649238586426, + "learning_rate": 0.00017599987521058216, + "loss": 0.1397, + "step": 23080 + }, + { + "epoch": 0.36017345729082173, + "grad_norm": 0.4577398896217346, + "learning_rate": 0.00017598947609242738, + "loss": 0.1072, + "step": 23090 + }, + { + "epoch": 0.36032944406314343, + "grad_norm": 0.6199226379394531, + "learning_rate": 0.0001759790769742726, + "loss": 0.3376, + "step": 23100 + }, + { + "epoch": 0.36048543083546514, + "grad_norm": 3.0997414588928223, + "learning_rate": 0.00017596867785611781, + "loss": 0.2746, + "step": 23110 + }, + { + "epoch": 0.36064141760778684, + "grad_norm": 1.6029644012451172, + "learning_rate": 0.00017595827873796303, + "loss": 0.1981, + "step": 23120 + }, + { + "epoch": 0.36079740438010854, + "grad_norm": 1.5090065002441406, + "learning_rate": 0.00017594787961980825, + "loss": 0.2736, + "step": 23130 + }, + { + "epoch": 0.36095339115243025, + "grad_norm": 2.2648723125457764, + "learning_rate": 0.00017593748050165347, + "loss": 0.2818, + "step": 23140 + }, + { + "epoch": 0.361109377924752, + "grad_norm": 2.141239643096924, + "learning_rate": 0.0001759270813834987, + "loss": 0.5181, + "step": 23150 + }, + { + "epoch": 0.3612653646970737, + "grad_norm": 0.3418712317943573, + "learning_rate": 0.0001759166822653439, + "loss": 0.1415, + "step": 23160 + }, + { + "epoch": 0.3614213514693954, + "grad_norm": 0.0884869173169136, + "learning_rate": 0.00017590628314718913, + "loss": 0.1143, + "step": 23170 + }, + { + "epoch": 0.3615773382417171, + "grad_norm": 1.6040232181549072, + "learning_rate": 0.00017589588402903435, + "loss": 0.2323, + "step": 23180 + }, + { + "epoch": 0.3617333250140388, + "grad_norm": 3.303593158721924, + "learning_rate": 0.00017588548491087957, + "loss": 0.4377, + "step": 23190 + }, + { + "epoch": 0.3618893117863605, + "grad_norm": 0.15572473406791687, + "learning_rate": 0.0001758750857927248, + "loss": 0.1258, + "step": 23200 + }, + { + "epoch": 0.3620452985586822, + "grad_norm": 1.3843340873718262, + "learning_rate": 0.00017586468667457, + "loss": 0.2149, + "step": 23210 + }, + { + "epoch": 0.36220128533100393, + "grad_norm": 2.793605089187622, + "learning_rate": 0.00017585428755641523, + "loss": 0.3255, + "step": 23220 + }, + { + "epoch": 0.36235727210332563, + "grad_norm": 0.739162266254425, + "learning_rate": 0.00017584388843826045, + "loss": 0.1594, + "step": 23230 + }, + { + "epoch": 0.36251325887564734, + "grad_norm": 2.422416925430298, + "learning_rate": 0.00017583348932010567, + "loss": 0.2542, + "step": 23240 + }, + { + "epoch": 0.36266924564796904, + "grad_norm": 3.132779359817505, + "learning_rate": 0.0001758230902019509, + "loss": 0.1944, + "step": 23250 + }, + { + "epoch": 0.36282523242029074, + "grad_norm": 1.8478645086288452, + "learning_rate": 0.0001758126910837961, + "loss": 0.3564, + "step": 23260 + }, + { + "epoch": 0.36298121919261245, + "grad_norm": 1.4015896320343018, + "learning_rate": 0.00017580229196564133, + "loss": 0.2156, + "step": 23270 + }, + { + "epoch": 0.36313720596493415, + "grad_norm": 0.81246018409729, + "learning_rate": 0.00017579189284748655, + "loss": 0.4223, + "step": 23280 + }, + { + "epoch": 0.36329319273725585, + "grad_norm": 2.741278886795044, + "learning_rate": 0.00017578149372933177, + "loss": 0.3829, + "step": 23290 + }, + { + "epoch": 0.3634491795095776, + "grad_norm": 5.9451680183410645, + "learning_rate": 0.000175771094611177, + "loss": 0.3809, + "step": 23300 + }, + { + "epoch": 0.3636051662818993, + "grad_norm": 2.035046100616455, + "learning_rate": 0.0001757606954930222, + "loss": 0.2685, + "step": 23310 + }, + { + "epoch": 0.363761153054221, + "grad_norm": 1.7791731357574463, + "learning_rate": 0.00017575029637486743, + "loss": 0.1609, + "step": 23320 + }, + { + "epoch": 0.3639171398265427, + "grad_norm": 2.3419806957244873, + "learning_rate": 0.00017573989725671265, + "loss": 0.2606, + "step": 23330 + }, + { + "epoch": 0.3640731265988644, + "grad_norm": 1.8289819955825806, + "learning_rate": 0.00017572949813855787, + "loss": 0.4495, + "step": 23340 + }, + { + "epoch": 0.36422911337118613, + "grad_norm": 0.7300289869308472, + "learning_rate": 0.00017571909902040309, + "loss": 0.3342, + "step": 23350 + }, + { + "epoch": 0.36438510014350783, + "grad_norm": 0.8906331062316895, + "learning_rate": 0.0001757086999022483, + "loss": 0.3509, + "step": 23360 + }, + { + "epoch": 0.36454108691582954, + "grad_norm": 1.325465440750122, + "learning_rate": 0.00017569830078409353, + "loss": 0.1627, + "step": 23370 + }, + { + "epoch": 0.36469707368815124, + "grad_norm": 3.0202317237854004, + "learning_rate": 0.00017568790166593874, + "loss": 0.3408, + "step": 23380 + }, + { + "epoch": 0.36485306046047294, + "grad_norm": 2.465944290161133, + "learning_rate": 0.00017567750254778394, + "loss": 0.2097, + "step": 23390 + }, + { + "epoch": 0.36500904723279465, + "grad_norm": 1.6739494800567627, + "learning_rate": 0.00017566710342962918, + "loss": 0.1482, + "step": 23400 + }, + { + "epoch": 0.36516503400511635, + "grad_norm": 0.5964159369468689, + "learning_rate": 0.00017565670431147438, + "loss": 0.1026, + "step": 23410 + }, + { + "epoch": 0.36532102077743805, + "grad_norm": 0.47202610969543457, + "learning_rate": 0.00017564630519331962, + "loss": 0.2756, + "step": 23420 + }, + { + "epoch": 0.36547700754975976, + "grad_norm": 2.9289498329162598, + "learning_rate": 0.00017563590607516482, + "loss": 0.2636, + "step": 23430 + }, + { + "epoch": 0.36563299432208146, + "grad_norm": 1.070677399635315, + "learning_rate": 0.00017562550695701006, + "loss": 0.1203, + "step": 23440 + }, + { + "epoch": 0.3657889810944032, + "grad_norm": 4.983691692352295, + "learning_rate": 0.00017561510783885526, + "loss": 0.3299, + "step": 23450 + }, + { + "epoch": 0.3659449678667249, + "grad_norm": 1.330994725227356, + "learning_rate": 0.0001756047087207005, + "loss": 0.1764, + "step": 23460 + }, + { + "epoch": 0.3661009546390466, + "grad_norm": 0.2447134107351303, + "learning_rate": 0.0001755943096025457, + "loss": 0.3493, + "step": 23470 + }, + { + "epoch": 0.36625694141136833, + "grad_norm": 0.8998173475265503, + "learning_rate": 0.00017558391048439094, + "loss": 0.2864, + "step": 23480 + }, + { + "epoch": 0.36641292818369003, + "grad_norm": 0.8895260691642761, + "learning_rate": 0.00017557351136623613, + "loss": 0.3781, + "step": 23490 + }, + { + "epoch": 0.36656891495601174, + "grad_norm": 2.5846433639526367, + "learning_rate": 0.00017556311224808138, + "loss": 0.1861, + "step": 23500 + }, + { + "epoch": 0.36672490172833344, + "grad_norm": 0.7542899250984192, + "learning_rate": 0.00017555271312992657, + "loss": 0.1388, + "step": 23510 + }, + { + "epoch": 0.36688088850065514, + "grad_norm": 2.791304349899292, + "learning_rate": 0.00017554231401177182, + "loss": 0.1454, + "step": 23520 + }, + { + "epoch": 0.36703687527297685, + "grad_norm": 2.2981035709381104, + "learning_rate": 0.000175531914893617, + "loss": 0.4003, + "step": 23530 + }, + { + "epoch": 0.36719286204529855, + "grad_norm": 0.9309369325637817, + "learning_rate": 0.00017552151577546226, + "loss": 0.3927, + "step": 23540 + }, + { + "epoch": 0.36734884881762025, + "grad_norm": 1.7445648908615112, + "learning_rate": 0.00017551111665730745, + "loss": 0.1801, + "step": 23550 + }, + { + "epoch": 0.36750483558994196, + "grad_norm": 0.42411547899246216, + "learning_rate": 0.0001755007175391527, + "loss": 0.2568, + "step": 23560 + }, + { + "epoch": 0.36766082236226366, + "grad_norm": 2.3300788402557373, + "learning_rate": 0.0001754903184209979, + "loss": 0.2386, + "step": 23570 + }, + { + "epoch": 0.36781680913458537, + "grad_norm": 3.840994358062744, + "learning_rate": 0.00017547991930284314, + "loss": 0.2798, + "step": 23580 + }, + { + "epoch": 0.36797279590690707, + "grad_norm": 2.637495279312134, + "learning_rate": 0.00017546952018468833, + "loss": 0.3753, + "step": 23590 + }, + { + "epoch": 0.3681287826792288, + "grad_norm": 1.5795278549194336, + "learning_rate": 0.00017545912106653358, + "loss": 0.3637, + "step": 23600 + }, + { + "epoch": 0.36828476945155053, + "grad_norm": 0.08003415167331696, + "learning_rate": 0.00017544872194837877, + "loss": 0.2208, + "step": 23610 + }, + { + "epoch": 0.36844075622387223, + "grad_norm": 0.6159747242927551, + "learning_rate": 0.00017543832283022402, + "loss": 0.133, + "step": 23620 + }, + { + "epoch": 0.36859674299619394, + "grad_norm": 1.2621487379074097, + "learning_rate": 0.0001754279237120692, + "loss": 0.3533, + "step": 23630 + }, + { + "epoch": 0.36875272976851564, + "grad_norm": 0.8740425109863281, + "learning_rate": 0.00017541752459391446, + "loss": 0.4002, + "step": 23640 + }, + { + "epoch": 0.36890871654083734, + "grad_norm": 0.23944754898548126, + "learning_rate": 0.00017540712547575965, + "loss": 0.3731, + "step": 23650 + }, + { + "epoch": 0.36906470331315905, + "grad_norm": 1.4508429765701294, + "learning_rate": 0.0001753967263576049, + "loss": 0.2836, + "step": 23660 + }, + { + "epoch": 0.36922069008548075, + "grad_norm": 0.8357475399971008, + "learning_rate": 0.0001753863272394501, + "loss": 0.3631, + "step": 23670 + }, + { + "epoch": 0.36937667685780246, + "grad_norm": 1.7004170417785645, + "learning_rate": 0.00017537592812129533, + "loss": 0.2985, + "step": 23680 + }, + { + "epoch": 0.36953266363012416, + "grad_norm": 1.9584474563598633, + "learning_rate": 0.00017536552900314053, + "loss": 0.2373, + "step": 23690 + }, + { + "epoch": 0.36968865040244586, + "grad_norm": 1.69345223903656, + "learning_rate": 0.00017535512988498577, + "loss": 0.6835, + "step": 23700 + }, + { + "epoch": 0.36984463717476757, + "grad_norm": 0.9576243162155151, + "learning_rate": 0.00017534473076683097, + "loss": 0.1574, + "step": 23710 + }, + { + "epoch": 0.37000062394708927, + "grad_norm": 3.3764142990112305, + "learning_rate": 0.0001753343316486762, + "loss": 0.2082, + "step": 23720 + }, + { + "epoch": 0.370156610719411, + "grad_norm": 2.110417604446411, + "learning_rate": 0.0001753239325305214, + "loss": 0.5735, + "step": 23730 + }, + { + "epoch": 0.3703125974917327, + "grad_norm": 3.3895039558410645, + "learning_rate": 0.00017531353341236665, + "loss": 0.3392, + "step": 23740 + }, + { + "epoch": 0.37046858426405443, + "grad_norm": 1.0715103149414062, + "learning_rate": 0.00017530313429421184, + "loss": 0.3691, + "step": 23750 + }, + { + "epoch": 0.37062457103637614, + "grad_norm": 0.3704458475112915, + "learning_rate": 0.0001752927351760571, + "loss": 0.4589, + "step": 23760 + }, + { + "epoch": 0.37078055780869784, + "grad_norm": 3.4379124641418457, + "learning_rate": 0.00017528233605790228, + "loss": 0.4224, + "step": 23770 + }, + { + "epoch": 0.37093654458101954, + "grad_norm": 1.3746329545974731, + "learning_rate": 0.00017527193693974753, + "loss": 0.3057, + "step": 23780 + }, + { + "epoch": 0.37109253135334125, + "grad_norm": 0.8195072412490845, + "learning_rate": 0.00017526153782159272, + "loss": 0.1931, + "step": 23790 + }, + { + "epoch": 0.37124851812566295, + "grad_norm": 0.6202637553215027, + "learning_rate": 0.00017525113870343797, + "loss": 0.3704, + "step": 23800 + }, + { + "epoch": 0.37140450489798466, + "grad_norm": 0.27026328444480896, + "learning_rate": 0.00017524073958528316, + "loss": 0.1713, + "step": 23810 + }, + { + "epoch": 0.37156049167030636, + "grad_norm": 1.6324976682662964, + "learning_rate": 0.0001752303404671284, + "loss": 0.3125, + "step": 23820 + }, + { + "epoch": 0.37171647844262806, + "grad_norm": 0.025506047531962395, + "learning_rate": 0.00017521994134897363, + "loss": 0.1637, + "step": 23830 + }, + { + "epoch": 0.37187246521494977, + "grad_norm": 4.021873474121094, + "learning_rate": 0.00017520954223081882, + "loss": 0.2982, + "step": 23840 + }, + { + "epoch": 0.37202845198727147, + "grad_norm": 1.620214819908142, + "learning_rate": 0.00017519914311266407, + "loss": 0.3241, + "step": 23850 + }, + { + "epoch": 0.3721844387595932, + "grad_norm": 2.8266336917877197, + "learning_rate": 0.00017518874399450926, + "loss": 0.4525, + "step": 23860 + }, + { + "epoch": 0.3723404255319149, + "grad_norm": 0.9824918508529663, + "learning_rate": 0.0001751783448763545, + "loss": 0.3745, + "step": 23870 + }, + { + "epoch": 0.3724964123042366, + "grad_norm": 0.8776328563690186, + "learning_rate": 0.0001751679457581997, + "loss": 0.2715, + "step": 23880 + }, + { + "epoch": 0.3726523990765583, + "grad_norm": 1.0546317100524902, + "learning_rate": 0.00017515754664004495, + "loss": 0.1969, + "step": 23890 + }, + { + "epoch": 0.37280838584888004, + "grad_norm": 1.7884589433670044, + "learning_rate": 0.00017514714752189014, + "loss": 0.1784, + "step": 23900 + }, + { + "epoch": 0.37296437262120175, + "grad_norm": 2.6703648567199707, + "learning_rate": 0.00017513674840373539, + "loss": 0.0942, + "step": 23910 + }, + { + "epoch": 0.37312035939352345, + "grad_norm": 1.487319827079773, + "learning_rate": 0.00017512634928558058, + "loss": 0.2205, + "step": 23920 + }, + { + "epoch": 0.37327634616584515, + "grad_norm": 1.2586092948913574, + "learning_rate": 0.00017511595016742582, + "loss": 0.1492, + "step": 23930 + }, + { + "epoch": 0.37343233293816686, + "grad_norm": 1.2256615161895752, + "learning_rate": 0.00017510555104927102, + "loss": 0.0919, + "step": 23940 + }, + { + "epoch": 0.37358831971048856, + "grad_norm": 0.0940847247838974, + "learning_rate": 0.00017509515193111626, + "loss": 0.291, + "step": 23950 + }, + { + "epoch": 0.37374430648281026, + "grad_norm": 2.2033092975616455, + "learning_rate": 0.00017508475281296146, + "loss": 0.2965, + "step": 23960 + }, + { + "epoch": 0.37390029325513197, + "grad_norm": 4.327813148498535, + "learning_rate": 0.0001750743536948067, + "loss": 0.5567, + "step": 23970 + }, + { + "epoch": 0.37405628002745367, + "grad_norm": 0.2586384117603302, + "learning_rate": 0.0001750639545766519, + "loss": 0.2756, + "step": 23980 + }, + { + "epoch": 0.3742122667997754, + "grad_norm": 1.7985506057739258, + "learning_rate": 0.00017505355545849714, + "loss": 0.2834, + "step": 23990 + }, + { + "epoch": 0.3743682535720971, + "grad_norm": 0.4995118975639343, + "learning_rate": 0.00017504315634034234, + "loss": 0.1613, + "step": 24000 + }, + { + "epoch": 0.3745242403444188, + "grad_norm": 0.8751158118247986, + "learning_rate": 0.00017503275722218758, + "loss": 0.2161, + "step": 24010 + }, + { + "epoch": 0.3746802271167405, + "grad_norm": 3.716893434524536, + "learning_rate": 0.00017502235810403277, + "loss": 0.2838, + "step": 24020 + }, + { + "epoch": 0.3748362138890622, + "grad_norm": 0.6846972107887268, + "learning_rate": 0.00017501195898587802, + "loss": 0.2838, + "step": 24030 + }, + { + "epoch": 0.3749922006613839, + "grad_norm": 4.110608100891113, + "learning_rate": 0.00017500155986772321, + "loss": 0.2258, + "step": 24040 + }, + { + "epoch": 0.37514818743370565, + "grad_norm": 2.093597888946533, + "learning_rate": 0.00017499116074956846, + "loss": 0.2144, + "step": 24050 + }, + { + "epoch": 0.37530417420602735, + "grad_norm": 0.34628939628601074, + "learning_rate": 0.00017498076163141365, + "loss": 0.1002, + "step": 24060 + }, + { + "epoch": 0.37546016097834906, + "grad_norm": 1.9365051984786987, + "learning_rate": 0.0001749703625132589, + "loss": 0.362, + "step": 24070 + }, + { + "epoch": 0.37561614775067076, + "grad_norm": 0.19635602831840515, + "learning_rate": 0.0001749599633951041, + "loss": 0.3215, + "step": 24080 + }, + { + "epoch": 0.37577213452299246, + "grad_norm": 0.7843225002288818, + "learning_rate": 0.00017494956427694934, + "loss": 0.1673, + "step": 24090 + }, + { + "epoch": 0.37592812129531417, + "grad_norm": 1.587742805480957, + "learning_rate": 0.00017493916515879453, + "loss": 0.3943, + "step": 24100 + }, + { + "epoch": 0.37608410806763587, + "grad_norm": 0.849687397480011, + "learning_rate": 0.00017492876604063978, + "loss": 0.1913, + "step": 24110 + }, + { + "epoch": 0.3762400948399576, + "grad_norm": 1.499902367591858, + "learning_rate": 0.00017491836692248497, + "loss": 0.2303, + "step": 24120 + }, + { + "epoch": 0.3763960816122793, + "grad_norm": 1.8464984893798828, + "learning_rate": 0.00017490796780433022, + "loss": 0.2977, + "step": 24130 + }, + { + "epoch": 0.376552068384601, + "grad_norm": 0.8660476803779602, + "learning_rate": 0.0001748975686861754, + "loss": 0.2752, + "step": 24140 + }, + { + "epoch": 0.3767080551569227, + "grad_norm": 0.30948498845100403, + "learning_rate": 0.00017488716956802066, + "loss": 0.219, + "step": 24150 + }, + { + "epoch": 0.3768640419292444, + "grad_norm": 0.37101125717163086, + "learning_rate": 0.00017487677044986585, + "loss": 0.3449, + "step": 24160 + }, + { + "epoch": 0.3770200287015661, + "grad_norm": 0.13139384984970093, + "learning_rate": 0.0001748663713317111, + "loss": 0.1291, + "step": 24170 + }, + { + "epoch": 0.3771760154738878, + "grad_norm": 1.7800439596176147, + "learning_rate": 0.0001748559722135563, + "loss": 0.3669, + "step": 24180 + }, + { + "epoch": 0.3773320022462095, + "grad_norm": 1.17131769657135, + "learning_rate": 0.00017484557309540154, + "loss": 0.2438, + "step": 24190 + }, + { + "epoch": 0.37748798901853126, + "grad_norm": 4.499425411224365, + "learning_rate": 0.00017483517397724673, + "loss": 0.4472, + "step": 24200 + }, + { + "epoch": 0.37764397579085296, + "grad_norm": 1.382828712463379, + "learning_rate": 0.00017482477485909197, + "loss": 0.3031, + "step": 24210 + }, + { + "epoch": 0.37779996256317466, + "grad_norm": 5.07600212097168, + "learning_rate": 0.00017481437574093717, + "loss": 0.14, + "step": 24220 + }, + { + "epoch": 0.37795594933549637, + "grad_norm": 2.3594703674316406, + "learning_rate": 0.00017480397662278241, + "loss": 0.1047, + "step": 24230 + }, + { + "epoch": 0.37811193610781807, + "grad_norm": 0.6317776441574097, + "learning_rate": 0.0001747935775046276, + "loss": 0.2673, + "step": 24240 + }, + { + "epoch": 0.3782679228801398, + "grad_norm": 0.2689322829246521, + "learning_rate": 0.00017478317838647285, + "loss": 0.4394, + "step": 24250 + }, + { + "epoch": 0.3784239096524615, + "grad_norm": 1.3794879913330078, + "learning_rate": 0.00017477277926831805, + "loss": 0.2043, + "step": 24260 + }, + { + "epoch": 0.3785798964247832, + "grad_norm": 2.3897125720977783, + "learning_rate": 0.0001747623801501633, + "loss": 0.3626, + "step": 24270 + }, + { + "epoch": 0.3787358831971049, + "grad_norm": 0.5998502373695374, + "learning_rate": 0.00017475198103200849, + "loss": 0.189, + "step": 24280 + }, + { + "epoch": 0.3788918699694266, + "grad_norm": 1.8097798824310303, + "learning_rate": 0.00017474158191385373, + "loss": 0.2584, + "step": 24290 + }, + { + "epoch": 0.3790478567417483, + "grad_norm": 0.2948211431503296, + "learning_rate": 0.00017473118279569892, + "loss": 0.2612, + "step": 24300 + }, + { + "epoch": 0.37920384351407, + "grad_norm": 3.582887887954712, + "learning_rate": 0.00017472078367754414, + "loss": 0.2086, + "step": 24310 + }, + { + "epoch": 0.3793598302863917, + "grad_norm": 0.9472853541374207, + "learning_rate": 0.00017471038455938936, + "loss": 0.2733, + "step": 24320 + }, + { + "epoch": 0.3795158170587134, + "grad_norm": 0.6683177947998047, + "learning_rate": 0.00017469998544123458, + "loss": 0.3209, + "step": 24330 + }, + { + "epoch": 0.3796718038310351, + "grad_norm": 0.8972749710083008, + "learning_rate": 0.0001746895863230798, + "loss": 0.2645, + "step": 24340 + }, + { + "epoch": 0.37982779060335686, + "grad_norm": 1.7964621782302856, + "learning_rate": 0.00017467918720492502, + "loss": 0.2823, + "step": 24350 + }, + { + "epoch": 0.37998377737567857, + "grad_norm": 0.6018429398536682, + "learning_rate": 0.00017466878808677024, + "loss": 0.1914, + "step": 24360 + }, + { + "epoch": 0.38013976414800027, + "grad_norm": 2.371281623840332, + "learning_rate": 0.00017465838896861546, + "loss": 0.242, + "step": 24370 + }, + { + "epoch": 0.380295750920322, + "grad_norm": 1.7452614307403564, + "learning_rate": 0.00017464798985046068, + "loss": 0.2579, + "step": 24380 + }, + { + "epoch": 0.3804517376926437, + "grad_norm": 0.49276208877563477, + "learning_rate": 0.0001746375907323059, + "loss": 0.3686, + "step": 24390 + }, + { + "epoch": 0.3806077244649654, + "grad_norm": 0.7708163261413574, + "learning_rate": 0.00017462719161415112, + "loss": 0.1572, + "step": 24400 + }, + { + "epoch": 0.3807637112372871, + "grad_norm": 2.104480028152466, + "learning_rate": 0.00017461679249599634, + "loss": 0.1402, + "step": 24410 + }, + { + "epoch": 0.3809196980096088, + "grad_norm": 1.6631516218185425, + "learning_rate": 0.00017460639337784156, + "loss": 0.2174, + "step": 24420 + }, + { + "epoch": 0.3810756847819305, + "grad_norm": 0.8534421920776367, + "learning_rate": 0.00017459599425968678, + "loss": 0.2835, + "step": 24430 + }, + { + "epoch": 0.3812316715542522, + "grad_norm": 2.7668111324310303, + "learning_rate": 0.000174585595141532, + "loss": 0.2767, + "step": 24440 + }, + { + "epoch": 0.3813876583265739, + "grad_norm": 0.2813465893268585, + "learning_rate": 0.00017457519602337722, + "loss": 0.2035, + "step": 24450 + }, + { + "epoch": 0.3815436450988956, + "grad_norm": 2.2833468914031982, + "learning_rate": 0.00017456479690522244, + "loss": 0.0935, + "step": 24460 + }, + { + "epoch": 0.3816996318712173, + "grad_norm": 2.791443109512329, + "learning_rate": 0.00017455439778706766, + "loss": 0.3437, + "step": 24470 + }, + { + "epoch": 0.381855618643539, + "grad_norm": 0.021708471700549126, + "learning_rate": 0.00017454399866891288, + "loss": 0.2596, + "step": 24480 + }, + { + "epoch": 0.3820116054158607, + "grad_norm": 0.723872184753418, + "learning_rate": 0.0001745335995507581, + "loss": 0.2104, + "step": 24490 + }, + { + "epoch": 0.3821675921881824, + "grad_norm": 1.3208684921264648, + "learning_rate": 0.00017452320043260332, + "loss": 0.2083, + "step": 24500 + }, + { + "epoch": 0.3823235789605042, + "grad_norm": 0.5394610166549683, + "learning_rate": 0.00017451280131444854, + "loss": 0.1535, + "step": 24510 + }, + { + "epoch": 0.3824795657328259, + "grad_norm": 0.7254071831703186, + "learning_rate": 0.00017450240219629376, + "loss": 0.1825, + "step": 24520 + }, + { + "epoch": 0.3826355525051476, + "grad_norm": 2.056793689727783, + "learning_rate": 0.00017449200307813898, + "loss": 0.1973, + "step": 24530 + }, + { + "epoch": 0.3827915392774693, + "grad_norm": 5.623049259185791, + "learning_rate": 0.0001744816039599842, + "loss": 0.1212, + "step": 24540 + }, + { + "epoch": 0.382947526049791, + "grad_norm": 2.166027069091797, + "learning_rate": 0.00017447120484182942, + "loss": 0.169, + "step": 24550 + }, + { + "epoch": 0.3831035128221127, + "grad_norm": 0.2908608913421631, + "learning_rate": 0.00017446080572367464, + "loss": 0.2275, + "step": 24560 + }, + { + "epoch": 0.3832594995944344, + "grad_norm": 4.945687770843506, + "learning_rate": 0.00017445040660551985, + "loss": 0.2189, + "step": 24570 + }, + { + "epoch": 0.3834154863667561, + "grad_norm": 0.8243374228477478, + "learning_rate": 0.00017444000748736507, + "loss": 0.3181, + "step": 24580 + }, + { + "epoch": 0.3835714731390778, + "grad_norm": 0.7564191818237305, + "learning_rate": 0.0001744296083692103, + "loss": 0.2197, + "step": 24590 + }, + { + "epoch": 0.3837274599113995, + "grad_norm": 0.5156879425048828, + "learning_rate": 0.00017441920925105551, + "loss": 0.3908, + "step": 24600 + }, + { + "epoch": 0.3838834466837212, + "grad_norm": 1.1277943849563599, + "learning_rate": 0.00017440881013290073, + "loss": 0.3542, + "step": 24610 + }, + { + "epoch": 0.3840394334560429, + "grad_norm": 1.3340795040130615, + "learning_rate": 0.00017439841101474595, + "loss": 0.2149, + "step": 24620 + }, + { + "epoch": 0.3841954202283646, + "grad_norm": 0.3093700110912323, + "learning_rate": 0.00017438801189659117, + "loss": 0.2087, + "step": 24630 + }, + { + "epoch": 0.3843514070006863, + "grad_norm": 0.2019505351781845, + "learning_rate": 0.0001743776127784364, + "loss": 0.1935, + "step": 24640 + }, + { + "epoch": 0.384507393773008, + "grad_norm": 1.4490982294082642, + "learning_rate": 0.0001743672136602816, + "loss": 0.2894, + "step": 24650 + }, + { + "epoch": 0.3846633805453298, + "grad_norm": 0.19713515043258667, + "learning_rate": 0.00017435681454212683, + "loss": 0.1492, + "step": 24660 + }, + { + "epoch": 0.3848193673176515, + "grad_norm": 0.8667004704475403, + "learning_rate": 0.00017434641542397205, + "loss": 0.2151, + "step": 24670 + }, + { + "epoch": 0.3849753540899732, + "grad_norm": 1.6552889347076416, + "learning_rate": 0.0001743360163058173, + "loss": 0.2266, + "step": 24680 + }, + { + "epoch": 0.3851313408622949, + "grad_norm": 2.130174398422241, + "learning_rate": 0.0001743256171876625, + "loss": 0.1835, + "step": 24690 + }, + { + "epoch": 0.3852873276346166, + "grad_norm": 10.104759216308594, + "learning_rate": 0.00017431521806950774, + "loss": 0.501, + "step": 24700 + }, + { + "epoch": 0.3854433144069383, + "grad_norm": 1.1952240467071533, + "learning_rate": 0.00017430481895135293, + "loss": 0.2539, + "step": 24710 + }, + { + "epoch": 0.38559930117926, + "grad_norm": 1.3514693975448608, + "learning_rate": 0.00017429441983319818, + "loss": 0.1838, + "step": 24720 + }, + { + "epoch": 0.3857552879515817, + "grad_norm": 3.540261745452881, + "learning_rate": 0.00017428402071504337, + "loss": 0.2005, + "step": 24730 + }, + { + "epoch": 0.3859112747239034, + "grad_norm": 0.3873755633831024, + "learning_rate": 0.00017427362159688862, + "loss": 0.5119, + "step": 24740 + }, + { + "epoch": 0.3860672614962251, + "grad_norm": 0.17622901499271393, + "learning_rate": 0.0001742632224787338, + "loss": 0.0946, + "step": 24750 + }, + { + "epoch": 0.3862232482685468, + "grad_norm": 1.4820303916931152, + "learning_rate": 0.00017425282336057903, + "loss": 0.1886, + "step": 24760 + }, + { + "epoch": 0.3863792350408685, + "grad_norm": 1.8511183261871338, + "learning_rate": 0.00017424242424242425, + "loss": 0.1238, + "step": 24770 + }, + { + "epoch": 0.3865352218131902, + "grad_norm": 1.7065273523330688, + "learning_rate": 0.00017423202512426947, + "loss": 0.383, + "step": 24780 + }, + { + "epoch": 0.3866912085855119, + "grad_norm": 0.050214044749736786, + "learning_rate": 0.0001742216260061147, + "loss": 0.2761, + "step": 24790 + }, + { + "epoch": 0.38684719535783363, + "grad_norm": 2.909736156463623, + "learning_rate": 0.0001742112268879599, + "loss": 0.26, + "step": 24800 + }, + { + "epoch": 0.3870031821301554, + "grad_norm": 4.363611221313477, + "learning_rate": 0.00017420082776980513, + "loss": 0.3548, + "step": 24810 + }, + { + "epoch": 0.3871591689024771, + "grad_norm": 0.3349769115447998, + "learning_rate": 0.00017419042865165035, + "loss": 0.3655, + "step": 24820 + }, + { + "epoch": 0.3873151556747988, + "grad_norm": 1.0107942819595337, + "learning_rate": 0.00017418002953349557, + "loss": 0.2663, + "step": 24830 + }, + { + "epoch": 0.3874711424471205, + "grad_norm": 2.1860833168029785, + "learning_rate": 0.00017416963041534079, + "loss": 0.231, + "step": 24840 + }, + { + "epoch": 0.3876271292194422, + "grad_norm": 8.581002235412598, + "learning_rate": 0.000174159231297186, + "loss": 0.5158, + "step": 24850 + }, + { + "epoch": 0.3877831159917639, + "grad_norm": 2.002713441848755, + "learning_rate": 0.00017414883217903122, + "loss": 0.1317, + "step": 24860 + }, + { + "epoch": 0.3879391027640856, + "grad_norm": 6.290743827819824, + "learning_rate": 0.00017413843306087644, + "loss": 0.5101, + "step": 24870 + }, + { + "epoch": 0.3880950895364073, + "grad_norm": 0.6607301831245422, + "learning_rate": 0.00017412803394272166, + "loss": 0.2735, + "step": 24880 + }, + { + "epoch": 0.388251076308729, + "grad_norm": 3.797792673110962, + "learning_rate": 0.00017411763482456688, + "loss": 0.3568, + "step": 24890 + }, + { + "epoch": 0.3884070630810507, + "grad_norm": 0.9680423140525818, + "learning_rate": 0.0001741072357064121, + "loss": 0.2093, + "step": 24900 + }, + { + "epoch": 0.3885630498533724, + "grad_norm": 0.7852384448051453, + "learning_rate": 0.00017409683658825732, + "loss": 0.2521, + "step": 24910 + }, + { + "epoch": 0.3887190366256941, + "grad_norm": 1.0753108263015747, + "learning_rate": 0.00017408643747010254, + "loss": 0.2128, + "step": 24920 + }, + { + "epoch": 0.38887502339801583, + "grad_norm": 0.5392903685569763, + "learning_rate": 0.00017407603835194776, + "loss": 0.3602, + "step": 24930 + }, + { + "epoch": 0.38903101017033753, + "grad_norm": 2.1266818046569824, + "learning_rate": 0.00017406563923379298, + "loss": 0.1904, + "step": 24940 + }, + { + "epoch": 0.38918699694265924, + "grad_norm": 2.467895984649658, + "learning_rate": 0.0001740552401156382, + "loss": 0.1976, + "step": 24950 + }, + { + "epoch": 0.389342983714981, + "grad_norm": 0.2751156985759735, + "learning_rate": 0.00017404484099748342, + "loss": 0.1306, + "step": 24960 + }, + { + "epoch": 0.3894989704873027, + "grad_norm": 1.5296525955200195, + "learning_rate": 0.00017403444187932864, + "loss": 0.2883, + "step": 24970 + }, + { + "epoch": 0.3896549572596244, + "grad_norm": 1.439092993736267, + "learning_rate": 0.00017402404276117386, + "loss": 0.5315, + "step": 24980 + }, + { + "epoch": 0.3898109440319461, + "grad_norm": 4.107110977172852, + "learning_rate": 0.00017401364364301908, + "loss": 0.3134, + "step": 24990 + }, + { + "epoch": 0.3899669308042678, + "grad_norm": 4.078581809997559, + "learning_rate": 0.0001740032445248643, + "loss": 0.3211, + "step": 25000 + }, + { + "epoch": 0.3901229175765895, + "grad_norm": 1.1098798513412476, + "learning_rate": 0.00017399284540670952, + "loss": 0.1202, + "step": 25010 + }, + { + "epoch": 0.3902789043489112, + "grad_norm": 0.5451555848121643, + "learning_rate": 0.00017398244628855474, + "loss": 0.1825, + "step": 25020 + }, + { + "epoch": 0.3904348911212329, + "grad_norm": 0.28062304854393005, + "learning_rate": 0.00017397204717039996, + "loss": 0.1119, + "step": 25030 + }, + { + "epoch": 0.3905908778935546, + "grad_norm": 0.23393197357654572, + "learning_rate": 0.00017396164805224518, + "loss": 0.1351, + "step": 25040 + }, + { + "epoch": 0.3907468646658763, + "grad_norm": 0.6696234345436096, + "learning_rate": 0.0001739512489340904, + "loss": 0.1619, + "step": 25050 + }, + { + "epoch": 0.39090285143819803, + "grad_norm": 1.2690047025680542, + "learning_rate": 0.00017394084981593562, + "loss": 0.1181, + "step": 25060 + }, + { + "epoch": 0.39105883821051973, + "grad_norm": 0.16582117974758148, + "learning_rate": 0.00017393045069778084, + "loss": 0.3847, + "step": 25070 + }, + { + "epoch": 0.39121482498284144, + "grad_norm": 2.0162479877471924, + "learning_rate": 0.00017392005157962606, + "loss": 0.1942, + "step": 25080 + }, + { + "epoch": 0.39137081175516314, + "grad_norm": 0.19528096914291382, + "learning_rate": 0.00017390965246147128, + "loss": 0.3542, + "step": 25090 + }, + { + "epoch": 0.39152679852748484, + "grad_norm": 3.472717046737671, + "learning_rate": 0.0001738992533433165, + "loss": 0.3962, + "step": 25100 + }, + { + "epoch": 0.3916827852998066, + "grad_norm": 4.016502857208252, + "learning_rate": 0.00017388885422516172, + "loss": 0.3952, + "step": 25110 + }, + { + "epoch": 0.3918387720721283, + "grad_norm": 2.819913625717163, + "learning_rate": 0.00017387845510700694, + "loss": 0.2647, + "step": 25120 + }, + { + "epoch": 0.39199475884445, + "grad_norm": 0.5896819233894348, + "learning_rate": 0.00017386805598885215, + "loss": 0.2561, + "step": 25130 + }, + { + "epoch": 0.3921507456167717, + "grad_norm": 0.09005920588970184, + "learning_rate": 0.00017385765687069737, + "loss": 0.2555, + "step": 25140 + }, + { + "epoch": 0.3923067323890934, + "grad_norm": 0.2739424407482147, + "learning_rate": 0.0001738472577525426, + "loss": 0.2043, + "step": 25150 + }, + { + "epoch": 0.3924627191614151, + "grad_norm": 0.3811953067779541, + "learning_rate": 0.00017383685863438781, + "loss": 0.1276, + "step": 25160 + }, + { + "epoch": 0.3926187059337368, + "grad_norm": 2.243910074234009, + "learning_rate": 0.00017382645951623303, + "loss": 0.1204, + "step": 25170 + }, + { + "epoch": 0.3927746927060585, + "grad_norm": 4.181073188781738, + "learning_rate": 0.00017381606039807825, + "loss": 0.3457, + "step": 25180 + }, + { + "epoch": 0.39293067947838023, + "grad_norm": 2.2200777530670166, + "learning_rate": 0.00017380566127992347, + "loss": 0.4306, + "step": 25190 + }, + { + "epoch": 0.39308666625070193, + "grad_norm": 1.0886428356170654, + "learning_rate": 0.0001737952621617687, + "loss": 0.1167, + "step": 25200 + }, + { + "epoch": 0.39324265302302364, + "grad_norm": 0.18910503387451172, + "learning_rate": 0.0001737848630436139, + "loss": 0.2225, + "step": 25210 + }, + { + "epoch": 0.39339863979534534, + "grad_norm": 6.567187786102295, + "learning_rate": 0.00017377446392545913, + "loss": 0.4123, + "step": 25220 + }, + { + "epoch": 0.39355462656766704, + "grad_norm": 0.9523065686225891, + "learning_rate": 0.00017376406480730435, + "loss": 0.4105, + "step": 25230 + }, + { + "epoch": 0.39371061333998875, + "grad_norm": 0.6237801313400269, + "learning_rate": 0.00017375366568914957, + "loss": 0.1869, + "step": 25240 + }, + { + "epoch": 0.39386660011231045, + "grad_norm": 2.5352957248687744, + "learning_rate": 0.0001737432665709948, + "loss": 0.2415, + "step": 25250 + }, + { + "epoch": 0.3940225868846322, + "grad_norm": 3.4460246562957764, + "learning_rate": 0.00017373286745284, + "loss": 0.5268, + "step": 25260 + }, + { + "epoch": 0.3941785736569539, + "grad_norm": 1.5610929727554321, + "learning_rate": 0.00017372246833468523, + "loss": 0.2573, + "step": 25270 + }, + { + "epoch": 0.3943345604292756, + "grad_norm": 0.4125531315803528, + "learning_rate": 0.00017371206921653045, + "loss": 0.2184, + "step": 25280 + }, + { + "epoch": 0.3944905472015973, + "grad_norm": 1.1488511562347412, + "learning_rate": 0.00017370167009837567, + "loss": 0.4354, + "step": 25290 + }, + { + "epoch": 0.394646533973919, + "grad_norm": 1.4139833450317383, + "learning_rate": 0.0001736912709802209, + "loss": 0.1193, + "step": 25300 + }, + { + "epoch": 0.3948025207462407, + "grad_norm": 1.4756202697753906, + "learning_rate": 0.0001736808718620661, + "loss": 0.2592, + "step": 25310 + }, + { + "epoch": 0.39495850751856243, + "grad_norm": 0.8054773807525635, + "learning_rate": 0.00017367047274391133, + "loss": 0.1781, + "step": 25320 + }, + { + "epoch": 0.39511449429088413, + "grad_norm": 0.814337432384491, + "learning_rate": 0.00017366007362575655, + "loss": 0.2944, + "step": 25330 + }, + { + "epoch": 0.39527048106320584, + "grad_norm": 1.5565993785858154, + "learning_rate": 0.00017364967450760177, + "loss": 0.1719, + "step": 25340 + }, + { + "epoch": 0.39542646783552754, + "grad_norm": 1.0450258255004883, + "learning_rate": 0.000173639275389447, + "loss": 0.1468, + "step": 25350 + }, + { + "epoch": 0.39558245460784924, + "grad_norm": 1.8846546411514282, + "learning_rate": 0.0001736288762712922, + "loss": 0.3879, + "step": 25360 + }, + { + "epoch": 0.39573844138017095, + "grad_norm": 2.735123634338379, + "learning_rate": 0.00017361847715313743, + "loss": 0.3024, + "step": 25370 + }, + { + "epoch": 0.39589442815249265, + "grad_norm": 1.1219899654388428, + "learning_rate": 0.00017360807803498265, + "loss": 0.1738, + "step": 25380 + }, + { + "epoch": 0.39605041492481435, + "grad_norm": 1.0913885831832886, + "learning_rate": 0.00017359767891682787, + "loss": 0.3033, + "step": 25390 + }, + { + "epoch": 0.39620640169713606, + "grad_norm": 4.076601028442383, + "learning_rate": 0.00017358727979867309, + "loss": 0.1912, + "step": 25400 + }, + { + "epoch": 0.3963623884694578, + "grad_norm": 2.658020496368408, + "learning_rate": 0.0001735768806805183, + "loss": 0.4013, + "step": 25410 + }, + { + "epoch": 0.3965183752417795, + "grad_norm": 1.6182650327682495, + "learning_rate": 0.00017356648156236352, + "loss": 0.1843, + "step": 25420 + }, + { + "epoch": 0.3966743620141012, + "grad_norm": 1.164387583732605, + "learning_rate": 0.00017355608244420874, + "loss": 0.5032, + "step": 25430 + }, + { + "epoch": 0.3968303487864229, + "grad_norm": 2.3382914066314697, + "learning_rate": 0.00017354568332605396, + "loss": 0.2038, + "step": 25440 + }, + { + "epoch": 0.39698633555874463, + "grad_norm": 0.6422469615936279, + "learning_rate": 0.00017353528420789918, + "loss": 0.3309, + "step": 25450 + }, + { + "epoch": 0.39714232233106633, + "grad_norm": 0.10722719877958298, + "learning_rate": 0.0001735248850897444, + "loss": 0.1669, + "step": 25460 + }, + { + "epoch": 0.39729830910338804, + "grad_norm": 2.376823663711548, + "learning_rate": 0.00017351448597158962, + "loss": 0.1205, + "step": 25470 + }, + { + "epoch": 0.39745429587570974, + "grad_norm": 2.6128294467926025, + "learning_rate": 0.00017350408685343484, + "loss": 0.308, + "step": 25480 + }, + { + "epoch": 0.39761028264803144, + "grad_norm": 7.0653839111328125, + "learning_rate": 0.00017349368773528006, + "loss": 0.335, + "step": 25490 + }, + { + "epoch": 0.39776626942035315, + "grad_norm": 3.4319369792938232, + "learning_rate": 0.00017348328861712528, + "loss": 0.2437, + "step": 25500 + }, + { + "epoch": 0.39792225619267485, + "grad_norm": 0.316787451505661, + "learning_rate": 0.0001734728894989705, + "loss": 0.1528, + "step": 25510 + }, + { + "epoch": 0.39807824296499655, + "grad_norm": 0.3672729432582855, + "learning_rate": 0.00017346249038081572, + "loss": 0.2575, + "step": 25520 + }, + { + "epoch": 0.39823422973731826, + "grad_norm": 0.002047836547717452, + "learning_rate": 0.00017345209126266094, + "loss": 0.188, + "step": 25530 + }, + { + "epoch": 0.39839021650963996, + "grad_norm": 0.35844242572784424, + "learning_rate": 0.00017344169214450616, + "loss": 0.3222, + "step": 25540 + }, + { + "epoch": 0.39854620328196166, + "grad_norm": 2.5270419120788574, + "learning_rate": 0.00017343129302635138, + "loss": 0.3519, + "step": 25550 + }, + { + "epoch": 0.3987021900542834, + "grad_norm": 1.8756089210510254, + "learning_rate": 0.0001734208939081966, + "loss": 0.1885, + "step": 25560 + }, + { + "epoch": 0.3988581768266051, + "grad_norm": 2.1598994731903076, + "learning_rate": 0.00017341049479004182, + "loss": 0.1869, + "step": 25570 + }, + { + "epoch": 0.39901416359892683, + "grad_norm": 3.2998921871185303, + "learning_rate": 0.00017340009567188704, + "loss": 0.2656, + "step": 25580 + }, + { + "epoch": 0.39917015037124853, + "grad_norm": 1.7846691608428955, + "learning_rate": 0.00017338969655373226, + "loss": 0.1338, + "step": 25590 + }, + { + "epoch": 0.39932613714357024, + "grad_norm": 1.244472861289978, + "learning_rate": 0.00017337929743557748, + "loss": 0.1339, + "step": 25600 + }, + { + "epoch": 0.39948212391589194, + "grad_norm": 0.9301228523254395, + "learning_rate": 0.0001733688983174227, + "loss": 0.2798, + "step": 25610 + }, + { + "epoch": 0.39963811068821364, + "grad_norm": 0.4158637523651123, + "learning_rate": 0.00017335849919926792, + "loss": 0.2263, + "step": 25620 + }, + { + "epoch": 0.39979409746053535, + "grad_norm": 0.22520148754119873, + "learning_rate": 0.00017334810008111314, + "loss": 0.318, + "step": 25630 + }, + { + "epoch": 0.39995008423285705, + "grad_norm": 0.23559512197971344, + "learning_rate": 0.00017333770096295836, + "loss": 0.3538, + "step": 25640 + }, + { + "epoch": 0.40010607100517875, + "grad_norm": 0.20354199409484863, + "learning_rate": 0.00017332730184480358, + "loss": 0.1814, + "step": 25650 + }, + { + "epoch": 0.40026205777750046, + "grad_norm": 0.8671756982803345, + "learning_rate": 0.00017331690272664877, + "loss": 0.3148, + "step": 25660 + }, + { + "epoch": 0.40041804454982216, + "grad_norm": 0.8490355014801025, + "learning_rate": 0.00017330650360849402, + "loss": 0.4086, + "step": 25670 + }, + { + "epoch": 0.40057403132214386, + "grad_norm": 1.5007946491241455, + "learning_rate": 0.0001732961044903392, + "loss": 0.2353, + "step": 25680 + }, + { + "epoch": 0.40073001809446557, + "grad_norm": 0.845577597618103, + "learning_rate": 0.00017328570537218445, + "loss": 0.3867, + "step": 25690 + }, + { + "epoch": 0.40088600486678727, + "grad_norm": 1.0520473718643188, + "learning_rate": 0.00017327530625402965, + "loss": 0.1485, + "step": 25700 + }, + { + "epoch": 0.40104199163910903, + "grad_norm": 2.261270761489868, + "learning_rate": 0.0001732649071358749, + "loss": 0.4676, + "step": 25710 + }, + { + "epoch": 0.40119797841143073, + "grad_norm": 0.5821919441223145, + "learning_rate": 0.0001732545080177201, + "loss": 0.2988, + "step": 25720 + }, + { + "epoch": 0.40135396518375244, + "grad_norm": 0.058162033557891846, + "learning_rate": 0.00017324410889956533, + "loss": 0.1443, + "step": 25730 + }, + { + "epoch": 0.40150995195607414, + "grad_norm": 0.879135012626648, + "learning_rate": 0.00017323370978141053, + "loss": 0.1521, + "step": 25740 + }, + { + "epoch": 0.40166593872839584, + "grad_norm": 1.554321050643921, + "learning_rate": 0.00017322331066325577, + "loss": 0.2826, + "step": 25750 + }, + { + "epoch": 0.40182192550071755, + "grad_norm": 1.952485203742981, + "learning_rate": 0.00017321291154510097, + "loss": 0.3709, + "step": 25760 + }, + { + "epoch": 0.40197791227303925, + "grad_norm": 1.2339608669281006, + "learning_rate": 0.0001732025124269462, + "loss": 0.1866, + "step": 25770 + }, + { + "epoch": 0.40213389904536095, + "grad_norm": 1.088102102279663, + "learning_rate": 0.0001731921133087914, + "loss": 0.2102, + "step": 25780 + }, + { + "epoch": 0.40228988581768266, + "grad_norm": 2.370701789855957, + "learning_rate": 0.00017318171419063665, + "loss": 0.483, + "step": 25790 + }, + { + "epoch": 0.40244587259000436, + "grad_norm": 0.45199036598205566, + "learning_rate": 0.00017317131507248184, + "loss": 0.2703, + "step": 25800 + }, + { + "epoch": 0.40260185936232606, + "grad_norm": 0.03911962732672691, + "learning_rate": 0.0001731609159543271, + "loss": 0.1925, + "step": 25810 + }, + { + "epoch": 0.40275784613464777, + "grad_norm": 0.7455926537513733, + "learning_rate": 0.00017315051683617228, + "loss": 0.322, + "step": 25820 + }, + { + "epoch": 0.40291383290696947, + "grad_norm": 0.1380135864019394, + "learning_rate": 0.00017314011771801753, + "loss": 0.2744, + "step": 25830 + }, + { + "epoch": 0.4030698196792912, + "grad_norm": 0.19523049890995026, + "learning_rate": 0.00017312971859986272, + "loss": 0.1108, + "step": 25840 + }, + { + "epoch": 0.4032258064516129, + "grad_norm": 6.985780715942383, + "learning_rate": 0.00017311931948170797, + "loss": 0.2474, + "step": 25850 + }, + { + "epoch": 0.40338179322393464, + "grad_norm": 0.7678770422935486, + "learning_rate": 0.00017310892036355316, + "loss": 0.5253, + "step": 25860 + }, + { + "epoch": 0.40353777999625634, + "grad_norm": 6.039801120758057, + "learning_rate": 0.0001730985212453984, + "loss": 0.3153, + "step": 25870 + }, + { + "epoch": 0.40369376676857804, + "grad_norm": 1.629223108291626, + "learning_rate": 0.0001730881221272436, + "loss": 0.3517, + "step": 25880 + }, + { + "epoch": 0.40384975354089975, + "grad_norm": 0.11092708259820938, + "learning_rate": 0.00017307772300908885, + "loss": 0.3185, + "step": 25890 + }, + { + "epoch": 0.40400574031322145, + "grad_norm": 0.4011111855506897, + "learning_rate": 0.00017306732389093404, + "loss": 0.1784, + "step": 25900 + }, + { + "epoch": 0.40416172708554315, + "grad_norm": 6.482341766357422, + "learning_rate": 0.0001730569247727793, + "loss": 0.3118, + "step": 25910 + }, + { + "epoch": 0.40431771385786486, + "grad_norm": 1.038848638534546, + "learning_rate": 0.00017304652565462448, + "loss": 0.2193, + "step": 25920 + }, + { + "epoch": 0.40447370063018656, + "grad_norm": 0.39992794394493103, + "learning_rate": 0.00017303612653646973, + "loss": 0.2276, + "step": 25930 + }, + { + "epoch": 0.40462968740250826, + "grad_norm": 0.17920106649398804, + "learning_rate": 0.00017302572741831492, + "loss": 0.2746, + "step": 25940 + }, + { + "epoch": 0.40478567417482997, + "grad_norm": 1.0355992317199707, + "learning_rate": 0.00017301532830016017, + "loss": 0.2802, + "step": 25950 + }, + { + "epoch": 0.40494166094715167, + "grad_norm": 0.6587734222412109, + "learning_rate": 0.00017300492918200536, + "loss": 0.3318, + "step": 25960 + }, + { + "epoch": 0.4050976477194734, + "grad_norm": 3.5396811962127686, + "learning_rate": 0.0001729945300638506, + "loss": 0.126, + "step": 25970 + }, + { + "epoch": 0.4052536344917951, + "grad_norm": 2.0736632347106934, + "learning_rate": 0.0001729841309456958, + "loss": 0.3545, + "step": 25980 + }, + { + "epoch": 0.4054096212641168, + "grad_norm": 1.2802982330322266, + "learning_rate": 0.00017297373182754104, + "loss": 0.3887, + "step": 25990 + }, + { + "epoch": 0.4055656080364385, + "grad_norm": 1.0736682415008545, + "learning_rate": 0.00017296333270938624, + "loss": 0.2459, + "step": 26000 + }, + { + "epoch": 0.40572159480876024, + "grad_norm": 0.07502292096614838, + "learning_rate": 0.00017295293359123148, + "loss": 0.2117, + "step": 26010 + }, + { + "epoch": 0.40587758158108195, + "grad_norm": 1.231829285621643, + "learning_rate": 0.0001729425344730767, + "loss": 0.1615, + "step": 26020 + }, + { + "epoch": 0.40603356835340365, + "grad_norm": 0.5829067826271057, + "learning_rate": 0.00017293213535492192, + "loss": 0.3034, + "step": 26030 + }, + { + "epoch": 0.40618955512572535, + "grad_norm": 0.7279021739959717, + "learning_rate": 0.00017292173623676714, + "loss": 0.2693, + "step": 26040 + }, + { + "epoch": 0.40634554189804706, + "grad_norm": 2.724191427230835, + "learning_rate": 0.00017291133711861236, + "loss": 0.2818, + "step": 26050 + }, + { + "epoch": 0.40650152867036876, + "grad_norm": 1.985594391822815, + "learning_rate": 0.00017290093800045758, + "loss": 0.1939, + "step": 26060 + }, + { + "epoch": 0.40665751544269046, + "grad_norm": 0.32234323024749756, + "learning_rate": 0.0001728905388823028, + "loss": 0.3899, + "step": 26070 + }, + { + "epoch": 0.40681350221501217, + "grad_norm": 1.9073020219802856, + "learning_rate": 0.00017288013976414802, + "loss": 0.2178, + "step": 26080 + }, + { + "epoch": 0.40696948898733387, + "grad_norm": 4.553755283355713, + "learning_rate": 0.00017286974064599324, + "loss": 0.2528, + "step": 26090 + }, + { + "epoch": 0.4071254757596556, + "grad_norm": 2.0203821659088135, + "learning_rate": 0.00017285934152783846, + "loss": 0.479, + "step": 26100 + }, + { + "epoch": 0.4072814625319773, + "grad_norm": 7.311489105224609, + "learning_rate": 0.00017284894240968365, + "loss": 0.3712, + "step": 26110 + }, + { + "epoch": 0.407437449304299, + "grad_norm": 0.31754928827285767, + "learning_rate": 0.0001728385432915289, + "loss": 0.1591, + "step": 26120 + }, + { + "epoch": 0.4075934360766207, + "grad_norm": 2.492265462875366, + "learning_rate": 0.0001728281441733741, + "loss": 0.2293, + "step": 26130 + }, + { + "epoch": 0.4077494228489424, + "grad_norm": 0.8722240328788757, + "learning_rate": 0.00017281774505521934, + "loss": 0.2089, + "step": 26140 + }, + { + "epoch": 0.4079054096212641, + "grad_norm": 0.9639162421226501, + "learning_rate": 0.00017280734593706453, + "loss": 0.198, + "step": 26150 + }, + { + "epoch": 0.40806139639358585, + "grad_norm": 3.1487390995025635, + "learning_rate": 0.00017279694681890978, + "loss": 0.3514, + "step": 26160 + }, + { + "epoch": 0.40821738316590755, + "grad_norm": 0.08169671893119812, + "learning_rate": 0.00017278654770075497, + "loss": 0.312, + "step": 26170 + }, + { + "epoch": 0.40837336993822926, + "grad_norm": 4.3061113357543945, + "learning_rate": 0.00017277614858260022, + "loss": 0.4154, + "step": 26180 + }, + { + "epoch": 0.40852935671055096, + "grad_norm": 0.76048344373703, + "learning_rate": 0.0001727657494644454, + "loss": 0.1986, + "step": 26190 + }, + { + "epoch": 0.40868534348287266, + "grad_norm": 1.2807050943374634, + "learning_rate": 0.00017275535034629066, + "loss": 0.2115, + "step": 26200 + }, + { + "epoch": 0.40884133025519437, + "grad_norm": 2.0285048484802246, + "learning_rate": 0.00017274495122813585, + "loss": 0.3973, + "step": 26210 + }, + { + "epoch": 0.40899731702751607, + "grad_norm": 2.566230297088623, + "learning_rate": 0.0001727345521099811, + "loss": 0.3299, + "step": 26220 + }, + { + "epoch": 0.4091533037998378, + "grad_norm": 3.5255863666534424, + "learning_rate": 0.0001727241529918263, + "loss": 0.318, + "step": 26230 + }, + { + "epoch": 0.4093092905721595, + "grad_norm": 1.0662899017333984, + "learning_rate": 0.00017271375387367153, + "loss": 0.2885, + "step": 26240 + }, + { + "epoch": 0.4094652773444812, + "grad_norm": 0.8745850324630737, + "learning_rate": 0.00017270335475551673, + "loss": 0.2976, + "step": 26250 + }, + { + "epoch": 0.4096212641168029, + "grad_norm": 0.9049080610275269, + "learning_rate": 0.00017269295563736197, + "loss": 0.3195, + "step": 26260 + }, + { + "epoch": 0.4097772508891246, + "grad_norm": 1.325914740562439, + "learning_rate": 0.00017268255651920717, + "loss": 0.0924, + "step": 26270 + }, + { + "epoch": 0.4099332376614463, + "grad_norm": 2.970973491668701, + "learning_rate": 0.0001726721574010524, + "loss": 0.1692, + "step": 26280 + }, + { + "epoch": 0.410089224433768, + "grad_norm": 1.4408884048461914, + "learning_rate": 0.0001726617582828976, + "loss": 0.431, + "step": 26290 + }, + { + "epoch": 0.4102452112060897, + "grad_norm": 0.04283915460109711, + "learning_rate": 0.00017265135916474285, + "loss": 0.2455, + "step": 26300 + }, + { + "epoch": 0.41040119797841146, + "grad_norm": 1.0744636058807373, + "learning_rate": 0.00017264096004658805, + "loss": 0.2285, + "step": 26310 + }, + { + "epoch": 0.41055718475073316, + "grad_norm": 1.8739361763000488, + "learning_rate": 0.0001726305609284333, + "loss": 0.2787, + "step": 26320 + }, + { + "epoch": 0.41071317152305487, + "grad_norm": 1.7582730054855347, + "learning_rate": 0.00017262016181027848, + "loss": 0.1584, + "step": 26330 + }, + { + "epoch": 0.41086915829537657, + "grad_norm": 0.7023983597755432, + "learning_rate": 0.00017260976269212373, + "loss": 0.141, + "step": 26340 + }, + { + "epoch": 0.41102514506769827, + "grad_norm": 0.15711884200572968, + "learning_rate": 0.00017259936357396892, + "loss": 0.2036, + "step": 26350 + }, + { + "epoch": 0.41118113184002, + "grad_norm": 0.7678493857383728, + "learning_rate": 0.00017258896445581417, + "loss": 0.0874, + "step": 26360 + }, + { + "epoch": 0.4113371186123417, + "grad_norm": 0.28819629549980164, + "learning_rate": 0.00017257856533765936, + "loss": 0.0755, + "step": 26370 + }, + { + "epoch": 0.4114931053846634, + "grad_norm": 3.942004919052124, + "learning_rate": 0.0001725681662195046, + "loss": 0.3809, + "step": 26380 + }, + { + "epoch": 0.4116490921569851, + "grad_norm": 1.877961277961731, + "learning_rate": 0.0001725577671013498, + "loss": 0.2686, + "step": 26390 + }, + { + "epoch": 0.4118050789293068, + "grad_norm": 2.786738157272339, + "learning_rate": 0.00017254736798319505, + "loss": 0.2331, + "step": 26400 + }, + { + "epoch": 0.4119610657016285, + "grad_norm": 1.9256407022476196, + "learning_rate": 0.00017253696886504024, + "loss": 0.199, + "step": 26410 + }, + { + "epoch": 0.4121170524739502, + "grad_norm": 0.2602331042289734, + "learning_rate": 0.0001725265697468855, + "loss": 0.234, + "step": 26420 + }, + { + "epoch": 0.4122730392462719, + "grad_norm": 0.18174906075000763, + "learning_rate": 0.00017251617062873068, + "loss": 0.3001, + "step": 26430 + }, + { + "epoch": 0.4124290260185936, + "grad_norm": 0.5017804503440857, + "learning_rate": 0.00017250577151057593, + "loss": 0.1645, + "step": 26440 + }, + { + "epoch": 0.4125850127909153, + "grad_norm": 0.4763779938220978, + "learning_rate": 0.00017249537239242112, + "loss": 0.2934, + "step": 26450 + }, + { + "epoch": 0.412740999563237, + "grad_norm": 2.08585524559021, + "learning_rate": 0.00017248497327426637, + "loss": 0.2201, + "step": 26460 + }, + { + "epoch": 0.41289698633555877, + "grad_norm": 4.861043453216553, + "learning_rate": 0.00017247457415611156, + "loss": 0.1817, + "step": 26470 + }, + { + "epoch": 0.4130529731078805, + "grad_norm": 0.5364469289779663, + "learning_rate": 0.0001724641750379568, + "loss": 0.2718, + "step": 26480 + }, + { + "epoch": 0.4132089598802022, + "grad_norm": 5.866112232208252, + "learning_rate": 0.000172453775919802, + "loss": 0.2052, + "step": 26490 + }, + { + "epoch": 0.4133649466525239, + "grad_norm": 2.871535539627075, + "learning_rate": 0.00017244337680164725, + "loss": 0.6411, + "step": 26500 + }, + { + "epoch": 0.4135209334248456, + "grad_norm": 1.2537444829940796, + "learning_rate": 0.00017243297768349244, + "loss": 0.2207, + "step": 26510 + }, + { + "epoch": 0.4136769201971673, + "grad_norm": 2.225438356399536, + "learning_rate": 0.00017242257856533768, + "loss": 0.1273, + "step": 26520 + }, + { + "epoch": 0.413832906969489, + "grad_norm": 2.568567991256714, + "learning_rate": 0.00017241217944718288, + "loss": 0.1398, + "step": 26530 + }, + { + "epoch": 0.4139888937418107, + "grad_norm": 2.5199825763702393, + "learning_rate": 0.00017240178032902812, + "loss": 0.1818, + "step": 26540 + }, + { + "epoch": 0.4141448805141324, + "grad_norm": 0.13436949253082275, + "learning_rate": 0.00017239138121087332, + "loss": 0.2353, + "step": 26550 + }, + { + "epoch": 0.4143008672864541, + "grad_norm": 0.5979320406913757, + "learning_rate": 0.00017238098209271854, + "loss": 0.1863, + "step": 26560 + }, + { + "epoch": 0.4144568540587758, + "grad_norm": 1.7884585857391357, + "learning_rate": 0.00017237058297456376, + "loss": 0.1638, + "step": 26570 + }, + { + "epoch": 0.4146128408310975, + "grad_norm": 0.7233499884605408, + "learning_rate": 0.00017236018385640898, + "loss": 0.241, + "step": 26580 + }, + { + "epoch": 0.4147688276034192, + "grad_norm": 0.3794673681259155, + "learning_rate": 0.0001723497847382542, + "loss": 0.4415, + "step": 26590 + }, + { + "epoch": 0.4149248143757409, + "grad_norm": 0.5616685748100281, + "learning_rate": 0.00017233938562009941, + "loss": 0.2108, + "step": 26600 + }, + { + "epoch": 0.4150808011480626, + "grad_norm": 1.6577153205871582, + "learning_rate": 0.00017232898650194463, + "loss": 0.2881, + "step": 26610 + }, + { + "epoch": 0.4152367879203844, + "grad_norm": 1.9585821628570557, + "learning_rate": 0.00017231858738378985, + "loss": 0.2909, + "step": 26620 + }, + { + "epoch": 0.4153927746927061, + "grad_norm": 0.20073258876800537, + "learning_rate": 0.00017230818826563507, + "loss": 0.271, + "step": 26630 + }, + { + "epoch": 0.4155487614650278, + "grad_norm": 1.5019569396972656, + "learning_rate": 0.0001722977891474803, + "loss": 0.1147, + "step": 26640 + }, + { + "epoch": 0.4157047482373495, + "grad_norm": 0.6672505736351013, + "learning_rate": 0.0001722873900293255, + "loss": 0.3165, + "step": 26650 + }, + { + "epoch": 0.4158607350096712, + "grad_norm": 1.6964812278747559, + "learning_rate": 0.00017227699091117073, + "loss": 0.3782, + "step": 26660 + }, + { + "epoch": 0.4160167217819929, + "grad_norm": 0.5775298476219177, + "learning_rate": 0.00017226659179301595, + "loss": 0.1922, + "step": 26670 + }, + { + "epoch": 0.4161727085543146, + "grad_norm": 0.8381676077842712, + "learning_rate": 0.00017225619267486117, + "loss": 0.2471, + "step": 26680 + }, + { + "epoch": 0.4163286953266363, + "grad_norm": 2.130768060684204, + "learning_rate": 0.0001722457935567064, + "loss": 0.2995, + "step": 26690 + }, + { + "epoch": 0.416484682098958, + "grad_norm": 3.691457509994507, + "learning_rate": 0.0001722353944385516, + "loss": 0.2249, + "step": 26700 + }, + { + "epoch": 0.4166406688712797, + "grad_norm": 1.0708967447280884, + "learning_rate": 0.00017222499532039683, + "loss": 0.2198, + "step": 26710 + }, + { + "epoch": 0.4167966556436014, + "grad_norm": 2.4624290466308594, + "learning_rate": 0.00017221459620224205, + "loss": 0.3118, + "step": 26720 + }, + { + "epoch": 0.4169526424159231, + "grad_norm": 2.53236985206604, + "learning_rate": 0.00017220419708408727, + "loss": 0.2879, + "step": 26730 + }, + { + "epoch": 0.4171086291882448, + "grad_norm": 2.291710376739502, + "learning_rate": 0.0001721937979659325, + "loss": 0.1504, + "step": 26740 + }, + { + "epoch": 0.4172646159605665, + "grad_norm": 1.2476613521575928, + "learning_rate": 0.0001721833988477777, + "loss": 0.1878, + "step": 26750 + }, + { + "epoch": 0.4174206027328882, + "grad_norm": 1.2871520519256592, + "learning_rate": 0.00017217299972962293, + "loss": 0.169, + "step": 26760 + }, + { + "epoch": 0.41757658950521, + "grad_norm": 0.6774186491966248, + "learning_rate": 0.00017216260061146815, + "loss": 0.1077, + "step": 26770 + }, + { + "epoch": 0.4177325762775317, + "grad_norm": 1.7373307943344116, + "learning_rate": 0.00017215220149331337, + "loss": 0.0733, + "step": 26780 + }, + { + "epoch": 0.4178885630498534, + "grad_norm": 1.529490351676941, + "learning_rate": 0.0001721418023751586, + "loss": 0.1776, + "step": 26790 + }, + { + "epoch": 0.4180445498221751, + "grad_norm": 0.0108403405174613, + "learning_rate": 0.0001721314032570038, + "loss": 0.1516, + "step": 26800 + }, + { + "epoch": 0.4182005365944968, + "grad_norm": 2.569559097290039, + "learning_rate": 0.00017212100413884903, + "loss": 0.4029, + "step": 26810 + }, + { + "epoch": 0.4183565233668185, + "grad_norm": 0.6180680394172668, + "learning_rate": 0.00017211060502069425, + "loss": 0.2863, + "step": 26820 + }, + { + "epoch": 0.4185125101391402, + "grad_norm": 3.9920032024383545, + "learning_rate": 0.00017210020590253947, + "loss": 0.362, + "step": 26830 + }, + { + "epoch": 0.4186684969114619, + "grad_norm": 0.04744652286171913, + "learning_rate": 0.00017208980678438469, + "loss": 0.2214, + "step": 26840 + }, + { + "epoch": 0.4188244836837836, + "grad_norm": 0.3316084146499634, + "learning_rate": 0.0001720794076662299, + "loss": 0.25, + "step": 26850 + }, + { + "epoch": 0.4189804704561053, + "grad_norm": 1.9856820106506348, + "learning_rate": 0.00017206900854807513, + "loss": 0.1854, + "step": 26860 + }, + { + "epoch": 0.419136457228427, + "grad_norm": 0.6092268228530884, + "learning_rate": 0.00017205860942992037, + "loss": 0.1751, + "step": 26870 + }, + { + "epoch": 0.4192924440007487, + "grad_norm": 0.550518274307251, + "learning_rate": 0.00017204821031176556, + "loss": 0.1249, + "step": 26880 + }, + { + "epoch": 0.4194484307730704, + "grad_norm": 3.0473105907440186, + "learning_rate": 0.0001720378111936108, + "loss": 0.2914, + "step": 26890 + }, + { + "epoch": 0.4196044175453921, + "grad_norm": 1.7360575199127197, + "learning_rate": 0.000172027412075456, + "loss": 0.2714, + "step": 26900 + }, + { + "epoch": 0.41976040431771383, + "grad_norm": 1.7520813941955566, + "learning_rate": 0.00017201701295730125, + "loss": 0.3297, + "step": 26910 + }, + { + "epoch": 0.4199163910900356, + "grad_norm": 1.1031064987182617, + "learning_rate": 0.00017200661383914644, + "loss": 0.1614, + "step": 26920 + }, + { + "epoch": 0.4200723778623573, + "grad_norm": 1.2025396823883057, + "learning_rate": 0.0001719962147209917, + "loss": 0.3231, + "step": 26930 + }, + { + "epoch": 0.420228364634679, + "grad_norm": 2.757190704345703, + "learning_rate": 0.00017198581560283688, + "loss": 0.2348, + "step": 26940 + }, + { + "epoch": 0.4203843514070007, + "grad_norm": 0.09398693591356277, + "learning_rate": 0.00017197541648468213, + "loss": 0.3347, + "step": 26950 + }, + { + "epoch": 0.4205403381793224, + "grad_norm": 0.4853222668170929, + "learning_rate": 0.00017196501736652732, + "loss": 0.1478, + "step": 26960 + }, + { + "epoch": 0.4206963249516441, + "grad_norm": 0.3310147821903229, + "learning_rate": 0.00017195461824837257, + "loss": 0.1509, + "step": 26970 + }, + { + "epoch": 0.4208523117239658, + "grad_norm": 3.506920337677002, + "learning_rate": 0.00017194421913021776, + "loss": 0.1917, + "step": 26980 + }, + { + "epoch": 0.4210082984962875, + "grad_norm": 3.7933475971221924, + "learning_rate": 0.000171933820012063, + "loss": 0.2472, + "step": 26990 + }, + { + "epoch": 0.4211642852686092, + "grad_norm": 0.7984077334403992, + "learning_rate": 0.0001719234208939082, + "loss": 0.2537, + "step": 27000 + }, + { + "epoch": 0.4213202720409309, + "grad_norm": 1.6348499059677124, + "learning_rate": 0.00017191302177575342, + "loss": 0.1245, + "step": 27010 + }, + { + "epoch": 0.4214762588132526, + "grad_norm": 0.02536233328282833, + "learning_rate": 0.00017190262265759864, + "loss": 0.2028, + "step": 27020 + }, + { + "epoch": 0.42163224558557433, + "grad_norm": 0.6552931070327759, + "learning_rate": 0.00017189222353944386, + "loss": 0.3634, + "step": 27030 + }, + { + "epoch": 0.42178823235789603, + "grad_norm": 2.812009334564209, + "learning_rate": 0.00017188182442128908, + "loss": 0.3509, + "step": 27040 + }, + { + "epoch": 0.42194421913021773, + "grad_norm": 4.050065994262695, + "learning_rate": 0.0001718714253031343, + "loss": 0.289, + "step": 27050 + }, + { + "epoch": 0.42210020590253944, + "grad_norm": 0.2938333749771118, + "learning_rate": 0.00017186102618497952, + "loss": 0.4355, + "step": 27060 + }, + { + "epoch": 0.4222561926748612, + "grad_norm": 3.093729257583618, + "learning_rate": 0.00017185062706682474, + "loss": 0.252, + "step": 27070 + }, + { + "epoch": 0.4224121794471829, + "grad_norm": 1.906206727027893, + "learning_rate": 0.00017184022794866996, + "loss": 0.2514, + "step": 27080 + }, + { + "epoch": 0.4225681662195046, + "grad_norm": 0.24559177458286285, + "learning_rate": 0.00017182982883051518, + "loss": 0.2819, + "step": 27090 + }, + { + "epoch": 0.4227241529918263, + "grad_norm": 2.733271598815918, + "learning_rate": 0.0001718194297123604, + "loss": 0.1515, + "step": 27100 + }, + { + "epoch": 0.422880139764148, + "grad_norm": 0.7841827869415283, + "learning_rate": 0.00017180903059420562, + "loss": 0.3472, + "step": 27110 + }, + { + "epoch": 0.4230361265364697, + "grad_norm": 2.464364528656006, + "learning_rate": 0.00017179863147605084, + "loss": 0.3206, + "step": 27120 + }, + { + "epoch": 0.4231921133087914, + "grad_norm": 1.816444993019104, + "learning_rate": 0.00017178823235789606, + "loss": 0.2256, + "step": 27130 + }, + { + "epoch": 0.4233481000811131, + "grad_norm": 0.4821169078350067, + "learning_rate": 0.00017177783323974128, + "loss": 0.2801, + "step": 27140 + }, + { + "epoch": 0.4235040868534348, + "grad_norm": 5.2493510246276855, + "learning_rate": 0.0001717674341215865, + "loss": 0.1718, + "step": 27150 + }, + { + "epoch": 0.42366007362575653, + "grad_norm": 2.033810615539551, + "learning_rate": 0.00017175703500343171, + "loss": 0.3248, + "step": 27160 + }, + { + "epoch": 0.42381606039807823, + "grad_norm": 1.5612436532974243, + "learning_rate": 0.00017174663588527693, + "loss": 0.1756, + "step": 27170 + }, + { + "epoch": 0.42397204717039993, + "grad_norm": 0.7761937975883484, + "learning_rate": 0.00017173623676712215, + "loss": 0.4597, + "step": 27180 + }, + { + "epoch": 0.42412803394272164, + "grad_norm": 1.778993844985962, + "learning_rate": 0.00017172583764896737, + "loss": 0.423, + "step": 27190 + }, + { + "epoch": 0.42428402071504334, + "grad_norm": 4.493278503417969, + "learning_rate": 0.0001717154385308126, + "loss": 0.2578, + "step": 27200 + }, + { + "epoch": 0.42444000748736505, + "grad_norm": 1.1676504611968994, + "learning_rate": 0.0001717050394126578, + "loss": 0.4056, + "step": 27210 + }, + { + "epoch": 0.4245959942596868, + "grad_norm": 2.2100558280944824, + "learning_rate": 0.00017169464029450303, + "loss": 0.4438, + "step": 27220 + }, + { + "epoch": 0.4247519810320085, + "grad_norm": 0.828824520111084, + "learning_rate": 0.00017168424117634825, + "loss": 0.223, + "step": 27230 + }, + { + "epoch": 0.4249079678043302, + "grad_norm": 0.8174467086791992, + "learning_rate": 0.00017167384205819347, + "loss": 0.3362, + "step": 27240 + }, + { + "epoch": 0.4250639545766519, + "grad_norm": 4.587996482849121, + "learning_rate": 0.0001716634429400387, + "loss": 0.322, + "step": 27250 + }, + { + "epoch": 0.4252199413489736, + "grad_norm": 0.6700587868690491, + "learning_rate": 0.0001716530438218839, + "loss": 0.1942, + "step": 27260 + }, + { + "epoch": 0.4253759281212953, + "grad_norm": 1.2546404600143433, + "learning_rate": 0.00017164264470372913, + "loss": 0.414, + "step": 27270 + }, + { + "epoch": 0.425531914893617, + "grad_norm": 1.416400671005249, + "learning_rate": 0.00017163224558557435, + "loss": 0.1427, + "step": 27280 + }, + { + "epoch": 0.42568790166593873, + "grad_norm": 5.11901330947876, + "learning_rate": 0.00017162184646741957, + "loss": 0.2341, + "step": 27290 + }, + { + "epoch": 0.42584388843826043, + "grad_norm": 3.462881326675415, + "learning_rate": 0.0001716114473492648, + "loss": 0.3235, + "step": 27300 + }, + { + "epoch": 0.42599987521058214, + "grad_norm": 0.9639891386032104, + "learning_rate": 0.00017160104823111, + "loss": 0.2178, + "step": 27310 + }, + { + "epoch": 0.42615586198290384, + "grad_norm": 1.8890912532806396, + "learning_rate": 0.00017159064911295523, + "loss": 0.1203, + "step": 27320 + }, + { + "epoch": 0.42631184875522554, + "grad_norm": 0.7909121513366699, + "learning_rate": 0.00017158024999480045, + "loss": 0.3414, + "step": 27330 + }, + { + "epoch": 0.42646783552754725, + "grad_norm": 0.3086511790752411, + "learning_rate": 0.00017156985087664567, + "loss": 0.3016, + "step": 27340 + }, + { + "epoch": 0.42662382229986895, + "grad_norm": 1.2022055387496948, + "learning_rate": 0.0001715594517584909, + "loss": 0.4312, + "step": 27350 + }, + { + "epoch": 0.42677980907219065, + "grad_norm": 0.8171715140342712, + "learning_rate": 0.0001715490526403361, + "loss": 0.4215, + "step": 27360 + }, + { + "epoch": 0.4269357958445124, + "grad_norm": 1.0555003881454468, + "learning_rate": 0.00017153865352218133, + "loss": 0.2145, + "step": 27370 + }, + { + "epoch": 0.4270917826168341, + "grad_norm": 0.567936360836029, + "learning_rate": 0.00017152825440402655, + "loss": 0.2033, + "step": 27380 + }, + { + "epoch": 0.4272477693891558, + "grad_norm": 0.9891570806503296, + "learning_rate": 0.00017151785528587177, + "loss": 0.12, + "step": 27390 + }, + { + "epoch": 0.4274037561614775, + "grad_norm": 0.8542338609695435, + "learning_rate": 0.00017150745616771699, + "loss": 0.3091, + "step": 27400 + }, + { + "epoch": 0.4275597429337992, + "grad_norm": 2.649017095565796, + "learning_rate": 0.0001714970570495622, + "loss": 0.3339, + "step": 27410 + }, + { + "epoch": 0.42771572970612093, + "grad_norm": 2.5795860290527344, + "learning_rate": 0.00017148665793140743, + "loss": 0.1787, + "step": 27420 + }, + { + "epoch": 0.42787171647844263, + "grad_norm": 1.3047258853912354, + "learning_rate": 0.00017147625881325265, + "loss": 0.2283, + "step": 27430 + }, + { + "epoch": 0.42802770325076434, + "grad_norm": 0.19904667139053345, + "learning_rate": 0.00017146585969509786, + "loss": 0.0928, + "step": 27440 + }, + { + "epoch": 0.42818369002308604, + "grad_norm": 0.3214978873729706, + "learning_rate": 0.00017145546057694308, + "loss": 0.2217, + "step": 27450 + }, + { + "epoch": 0.42833967679540774, + "grad_norm": 0.4364519417285919, + "learning_rate": 0.0001714450614587883, + "loss": 0.4297, + "step": 27460 + }, + { + "epoch": 0.42849566356772945, + "grad_norm": 1.0661823749542236, + "learning_rate": 0.00017143466234063352, + "loss": 0.2568, + "step": 27470 + }, + { + "epoch": 0.42865165034005115, + "grad_norm": 1.4220117330551147, + "learning_rate": 0.00017142426322247874, + "loss": 0.2036, + "step": 27480 + }, + { + "epoch": 0.42880763711237285, + "grad_norm": 0.5348262786865234, + "learning_rate": 0.00017141386410432396, + "loss": 0.2923, + "step": 27490 + }, + { + "epoch": 0.42896362388469456, + "grad_norm": 1.839390754699707, + "learning_rate": 0.00017140346498616918, + "loss": 0.269, + "step": 27500 + }, + { + "epoch": 0.42911961065701626, + "grad_norm": 0.6216630339622498, + "learning_rate": 0.0001713930658680144, + "loss": 0.2909, + "step": 27510 + }, + { + "epoch": 0.429275597429338, + "grad_norm": 0.5074413418769836, + "learning_rate": 0.00017138266674985962, + "loss": 0.3876, + "step": 27520 + }, + { + "epoch": 0.4294315842016597, + "grad_norm": 0.9923599362373352, + "learning_rate": 0.00017137226763170484, + "loss": 0.1267, + "step": 27530 + }, + { + "epoch": 0.4295875709739814, + "grad_norm": 0.39644038677215576, + "learning_rate": 0.00017136186851355006, + "loss": 0.269, + "step": 27540 + }, + { + "epoch": 0.42974355774630313, + "grad_norm": 1.7660781145095825, + "learning_rate": 0.00017135146939539528, + "loss": 0.2071, + "step": 27550 + }, + { + "epoch": 0.42989954451862483, + "grad_norm": 1.513165831565857, + "learning_rate": 0.0001713410702772405, + "loss": 0.3098, + "step": 27560 + }, + { + "epoch": 0.43005553129094654, + "grad_norm": 0.34492722153663635, + "learning_rate": 0.00017133067115908572, + "loss": 0.2922, + "step": 27570 + }, + { + "epoch": 0.43021151806326824, + "grad_norm": 0.7684240937232971, + "learning_rate": 0.00017132027204093094, + "loss": 0.4982, + "step": 27580 + }, + { + "epoch": 0.43036750483558994, + "grad_norm": 4.070146560668945, + "learning_rate": 0.00017130987292277616, + "loss": 0.201, + "step": 27590 + }, + { + "epoch": 0.43052349160791165, + "grad_norm": 0.21442419290542603, + "learning_rate": 0.00017129947380462138, + "loss": 0.1771, + "step": 27600 + }, + { + "epoch": 0.43067947838023335, + "grad_norm": 0.14202871918678284, + "learning_rate": 0.0001712890746864666, + "loss": 0.281, + "step": 27610 + }, + { + "epoch": 0.43083546515255505, + "grad_norm": 0.6332352757453918, + "learning_rate": 0.00017127867556831182, + "loss": 0.0819, + "step": 27620 + }, + { + "epoch": 0.43099145192487676, + "grad_norm": 5.952022552490234, + "learning_rate": 0.00017126827645015704, + "loss": 0.8225, + "step": 27630 + }, + { + "epoch": 0.43114743869719846, + "grad_norm": 1.9152021408081055, + "learning_rate": 0.00017125787733200226, + "loss": 0.258, + "step": 27640 + }, + { + "epoch": 0.43130342546952016, + "grad_norm": 1.3850319385528564, + "learning_rate": 0.00017124747821384748, + "loss": 0.2011, + "step": 27650 + }, + { + "epoch": 0.43145941224184187, + "grad_norm": 0.3383638560771942, + "learning_rate": 0.0001712370790956927, + "loss": 0.1751, + "step": 27660 + }, + { + "epoch": 0.4316153990141636, + "grad_norm": 1.781374216079712, + "learning_rate": 0.00017122667997753792, + "loss": 0.2253, + "step": 27670 + }, + { + "epoch": 0.43177138578648533, + "grad_norm": 1.7024178504943848, + "learning_rate": 0.00017121628085938314, + "loss": 0.5244, + "step": 27680 + }, + { + "epoch": 0.43192737255880703, + "grad_norm": 0.7735491394996643, + "learning_rate": 0.00017120588174122836, + "loss": 0.179, + "step": 27690 + }, + { + "epoch": 0.43208335933112874, + "grad_norm": 1.287466049194336, + "learning_rate": 0.00017119548262307358, + "loss": 0.2911, + "step": 27700 + }, + { + "epoch": 0.43223934610345044, + "grad_norm": 1.8219951391220093, + "learning_rate": 0.0001711850835049188, + "loss": 0.215, + "step": 27710 + }, + { + "epoch": 0.43239533287577214, + "grad_norm": 2.1505789756774902, + "learning_rate": 0.00017117468438676401, + "loss": 0.2211, + "step": 27720 + }, + { + "epoch": 0.43255131964809385, + "grad_norm": 1.0259547233581543, + "learning_rate": 0.00017116428526860923, + "loss": 0.2124, + "step": 27730 + }, + { + "epoch": 0.43270730642041555, + "grad_norm": 1.5184487104415894, + "learning_rate": 0.00017115388615045445, + "loss": 0.4171, + "step": 27740 + }, + { + "epoch": 0.43286329319273725, + "grad_norm": 3.1281065940856934, + "learning_rate": 0.00017114348703229967, + "loss": 0.2821, + "step": 27750 + }, + { + "epoch": 0.43301927996505896, + "grad_norm": 1.4081364870071411, + "learning_rate": 0.0001711330879141449, + "loss": 0.1274, + "step": 27760 + }, + { + "epoch": 0.43317526673738066, + "grad_norm": 1.411218285560608, + "learning_rate": 0.0001711226887959901, + "loss": 0.2503, + "step": 27770 + }, + { + "epoch": 0.43333125350970236, + "grad_norm": 0.954667329788208, + "learning_rate": 0.00017111228967783533, + "loss": 0.3665, + "step": 27780 + }, + { + "epoch": 0.43348724028202407, + "grad_norm": 2.8865349292755127, + "learning_rate": 0.00017110189055968055, + "loss": 0.1628, + "step": 27790 + }, + { + "epoch": 0.43364322705434577, + "grad_norm": 0.48402658104896545, + "learning_rate": 0.00017109149144152577, + "loss": 0.1422, + "step": 27800 + }, + { + "epoch": 0.4337992138266675, + "grad_norm": 1.419069766998291, + "learning_rate": 0.000171081092323371, + "loss": 0.0997, + "step": 27810 + }, + { + "epoch": 0.43395520059898923, + "grad_norm": 2.131889820098877, + "learning_rate": 0.0001710706932052162, + "loss": 0.3298, + "step": 27820 + }, + { + "epoch": 0.43411118737131094, + "grad_norm": 2.728317975997925, + "learning_rate": 0.00017106029408706143, + "loss": 0.279, + "step": 27830 + }, + { + "epoch": 0.43426717414363264, + "grad_norm": 0.8831292986869812, + "learning_rate": 0.00017104989496890665, + "loss": 0.2337, + "step": 27840 + }, + { + "epoch": 0.43442316091595434, + "grad_norm": 0.9926959276199341, + "learning_rate": 0.00017103949585075187, + "loss": 0.1986, + "step": 27850 + }, + { + "epoch": 0.43457914768827605, + "grad_norm": 0.5922303795814514, + "learning_rate": 0.0001710290967325971, + "loss": 0.1999, + "step": 27860 + }, + { + "epoch": 0.43473513446059775, + "grad_norm": 0.6125643253326416, + "learning_rate": 0.0001710186976144423, + "loss": 0.5852, + "step": 27870 + }, + { + "epoch": 0.43489112123291945, + "grad_norm": 0.8395240306854248, + "learning_rate": 0.00017100829849628753, + "loss": 0.2016, + "step": 27880 + }, + { + "epoch": 0.43504710800524116, + "grad_norm": 1.4695276021957397, + "learning_rate": 0.00017099789937813275, + "loss": 0.267, + "step": 27890 + }, + { + "epoch": 0.43520309477756286, + "grad_norm": 3.541914939880371, + "learning_rate": 0.00017098750025997797, + "loss": 0.2069, + "step": 27900 + }, + { + "epoch": 0.43535908154988456, + "grad_norm": 2.181877613067627, + "learning_rate": 0.00017097710114182316, + "loss": 0.4444, + "step": 27910 + }, + { + "epoch": 0.43551506832220627, + "grad_norm": 1.0567023754119873, + "learning_rate": 0.0001709667020236684, + "loss": 0.2815, + "step": 27920 + }, + { + "epoch": 0.43567105509452797, + "grad_norm": 1.270294427871704, + "learning_rate": 0.0001709563029055136, + "loss": 0.3011, + "step": 27930 + }, + { + "epoch": 0.4358270418668497, + "grad_norm": 0.1230715811252594, + "learning_rate": 0.00017094590378735885, + "loss": 0.1427, + "step": 27940 + }, + { + "epoch": 0.4359830286391714, + "grad_norm": 2.7168827056884766, + "learning_rate": 0.00017093550466920404, + "loss": 0.4045, + "step": 27950 + }, + { + "epoch": 0.4361390154114931, + "grad_norm": 2.149280548095703, + "learning_rate": 0.00017092510555104929, + "loss": 0.2767, + "step": 27960 + }, + { + "epoch": 0.43629500218381484, + "grad_norm": 3.3689825534820557, + "learning_rate": 0.00017091470643289448, + "loss": 0.5126, + "step": 27970 + }, + { + "epoch": 0.43645098895613654, + "grad_norm": 0.050475817173719406, + "learning_rate": 0.00017090430731473973, + "loss": 0.3973, + "step": 27980 + }, + { + "epoch": 0.43660697572845825, + "grad_norm": 0.14375102519989014, + "learning_rate": 0.00017089390819658492, + "loss": 0.3071, + "step": 27990 + }, + { + "epoch": 0.43676296250077995, + "grad_norm": 0.25722286105155945, + "learning_rate": 0.00017088350907843016, + "loss": 0.0992, + "step": 28000 + }, + { + "epoch": 0.43691894927310165, + "grad_norm": 0.9952245354652405, + "learning_rate": 0.00017087310996027536, + "loss": 0.3007, + "step": 28010 + }, + { + "epoch": 0.43707493604542336, + "grad_norm": 2.0607028007507324, + "learning_rate": 0.0001708627108421206, + "loss": 0.1433, + "step": 28020 + }, + { + "epoch": 0.43723092281774506, + "grad_norm": 3.218749523162842, + "learning_rate": 0.0001708523117239658, + "loss": 0.3536, + "step": 28030 + }, + { + "epoch": 0.43738690959006676, + "grad_norm": 0.21981604397296906, + "learning_rate": 0.00017084191260581104, + "loss": 0.238, + "step": 28040 + }, + { + "epoch": 0.43754289636238847, + "grad_norm": 1.2609288692474365, + "learning_rate": 0.00017083151348765624, + "loss": 0.3354, + "step": 28050 + }, + { + "epoch": 0.43769888313471017, + "grad_norm": 0.6399383544921875, + "learning_rate": 0.00017082111436950148, + "loss": 0.2271, + "step": 28060 + }, + { + "epoch": 0.4378548699070319, + "grad_norm": 3.4272584915161133, + "learning_rate": 0.00017081071525134667, + "loss": 0.3276, + "step": 28070 + }, + { + "epoch": 0.4380108566793536, + "grad_norm": 1.3369107246398926, + "learning_rate": 0.00017080031613319192, + "loss": 0.2195, + "step": 28080 + }, + { + "epoch": 0.4381668434516753, + "grad_norm": 3.4089694023132324, + "learning_rate": 0.00017078991701503711, + "loss": 0.2471, + "step": 28090 + }, + { + "epoch": 0.438322830223997, + "grad_norm": 0.04700610414147377, + "learning_rate": 0.00017077951789688236, + "loss": 0.3169, + "step": 28100 + }, + { + "epoch": 0.4384788169963187, + "grad_norm": 1.6034051179885864, + "learning_rate": 0.00017076911877872755, + "loss": 0.1929, + "step": 28110 + }, + { + "epoch": 0.43863480376864045, + "grad_norm": 1.5593754053115845, + "learning_rate": 0.0001707587196605728, + "loss": 0.1722, + "step": 28120 + }, + { + "epoch": 0.43879079054096215, + "grad_norm": 0.03455871343612671, + "learning_rate": 0.000170748320542418, + "loss": 0.1934, + "step": 28130 + }, + { + "epoch": 0.43894677731328385, + "grad_norm": 1.028444766998291, + "learning_rate": 0.00017073792142426324, + "loss": 0.3345, + "step": 28140 + }, + { + "epoch": 0.43910276408560556, + "grad_norm": 0.1876237541437149, + "learning_rate": 0.00017072752230610843, + "loss": 0.2392, + "step": 28150 + }, + { + "epoch": 0.43925875085792726, + "grad_norm": 1.046366810798645, + "learning_rate": 0.00017071712318795368, + "loss": 0.1491, + "step": 28160 + }, + { + "epoch": 0.43941473763024896, + "grad_norm": 1.3504830598831177, + "learning_rate": 0.00017070672406979887, + "loss": 0.192, + "step": 28170 + }, + { + "epoch": 0.43957072440257067, + "grad_norm": 1.6892421245574951, + "learning_rate": 0.00017069632495164412, + "loss": 0.291, + "step": 28180 + }, + { + "epoch": 0.43972671117489237, + "grad_norm": 0.6202647089958191, + "learning_rate": 0.0001706859258334893, + "loss": 0.3046, + "step": 28190 + }, + { + "epoch": 0.4398826979472141, + "grad_norm": 1.9218010902404785, + "learning_rate": 0.00017067552671533456, + "loss": 0.318, + "step": 28200 + }, + { + "epoch": 0.4400386847195358, + "grad_norm": 5.535128116607666, + "learning_rate": 0.00017066512759717978, + "loss": 0.5163, + "step": 28210 + }, + { + "epoch": 0.4401946714918575, + "grad_norm": 1.4575244188308716, + "learning_rate": 0.000170654728479025, + "loss": 0.468, + "step": 28220 + }, + { + "epoch": 0.4403506582641792, + "grad_norm": 0.457217812538147, + "learning_rate": 0.00017064432936087022, + "loss": 0.2568, + "step": 28230 + }, + { + "epoch": 0.4405066450365009, + "grad_norm": 1.087171196937561, + "learning_rate": 0.00017063393024271544, + "loss": 0.3977, + "step": 28240 + }, + { + "epoch": 0.4406626318088226, + "grad_norm": 0.060369931161403656, + "learning_rate": 0.00017062353112456066, + "loss": 0.1082, + "step": 28250 + }, + { + "epoch": 0.4408186185811443, + "grad_norm": 0.3021914064884186, + "learning_rate": 0.00017061313200640588, + "loss": 0.2022, + "step": 28260 + }, + { + "epoch": 0.44097460535346605, + "grad_norm": 1.7404426336288452, + "learning_rate": 0.0001706027328882511, + "loss": 0.3621, + "step": 28270 + }, + { + "epoch": 0.44113059212578776, + "grad_norm": 0.21145115792751312, + "learning_rate": 0.00017059233377009631, + "loss": 0.2772, + "step": 28280 + }, + { + "epoch": 0.44128657889810946, + "grad_norm": 3.2288665771484375, + "learning_rate": 0.00017058193465194153, + "loss": 0.2915, + "step": 28290 + }, + { + "epoch": 0.44144256567043116, + "grad_norm": 0.24990910291671753, + "learning_rate": 0.00017057153553378675, + "loss": 0.1293, + "step": 28300 + }, + { + "epoch": 0.44159855244275287, + "grad_norm": 0.9315836429595947, + "learning_rate": 0.00017056113641563197, + "loss": 0.2318, + "step": 28310 + }, + { + "epoch": 0.44175453921507457, + "grad_norm": 2.955906391143799, + "learning_rate": 0.0001705507372974772, + "loss": 0.6262, + "step": 28320 + }, + { + "epoch": 0.4419105259873963, + "grad_norm": 4.838889122009277, + "learning_rate": 0.0001705403381793224, + "loss": 0.3053, + "step": 28330 + }, + { + "epoch": 0.442066512759718, + "grad_norm": 0.05758214741945267, + "learning_rate": 0.00017052993906116763, + "loss": 0.1608, + "step": 28340 + }, + { + "epoch": 0.4422224995320397, + "grad_norm": 0.030875004827976227, + "learning_rate": 0.00017051953994301285, + "loss": 0.0683, + "step": 28350 + }, + { + "epoch": 0.4423784863043614, + "grad_norm": 0.6623532176017761, + "learning_rate": 0.00017050914082485807, + "loss": 0.2798, + "step": 28360 + }, + { + "epoch": 0.4425344730766831, + "grad_norm": 1.7801457643508911, + "learning_rate": 0.0001704987417067033, + "loss": 0.3352, + "step": 28370 + }, + { + "epoch": 0.4426904598490048, + "grad_norm": 3.1258904933929443, + "learning_rate": 0.00017048834258854848, + "loss": 0.1466, + "step": 28380 + }, + { + "epoch": 0.4428464466213265, + "grad_norm": 2.6470370292663574, + "learning_rate": 0.00017047794347039373, + "loss": 0.1695, + "step": 28390 + }, + { + "epoch": 0.4430024333936482, + "grad_norm": 2.824866771697998, + "learning_rate": 0.00017046754435223892, + "loss": 0.1682, + "step": 28400 + }, + { + "epoch": 0.4431584201659699, + "grad_norm": 2.454986333847046, + "learning_rate": 0.00017045714523408417, + "loss": 0.3083, + "step": 28410 + }, + { + "epoch": 0.4433144069382916, + "grad_norm": 2.338167667388916, + "learning_rate": 0.00017044674611592936, + "loss": 0.2881, + "step": 28420 + }, + { + "epoch": 0.44347039371061336, + "grad_norm": 1.9540684223175049, + "learning_rate": 0.0001704363469977746, + "loss": 0.2409, + "step": 28430 + }, + { + "epoch": 0.44362638048293507, + "grad_norm": 0.14418472349643707, + "learning_rate": 0.0001704259478796198, + "loss": 0.343, + "step": 28440 + }, + { + "epoch": 0.44378236725525677, + "grad_norm": 3.184375047683716, + "learning_rate": 0.00017041554876146505, + "loss": 0.2373, + "step": 28450 + }, + { + "epoch": 0.4439383540275785, + "grad_norm": 0.17831261456012726, + "learning_rate": 0.00017040514964331024, + "loss": 0.213, + "step": 28460 + }, + { + "epoch": 0.4440943407999002, + "grad_norm": 2.2932331562042236, + "learning_rate": 0.0001703947505251555, + "loss": 0.3775, + "step": 28470 + }, + { + "epoch": 0.4442503275722219, + "grad_norm": 1.882422924041748, + "learning_rate": 0.00017038435140700068, + "loss": 0.2646, + "step": 28480 + }, + { + "epoch": 0.4444063143445436, + "grad_norm": 0.27750319242477417, + "learning_rate": 0.00017037395228884593, + "loss": 0.2114, + "step": 28490 + }, + { + "epoch": 0.4445623011168653, + "grad_norm": 0.996457040309906, + "learning_rate": 0.00017036355317069112, + "loss": 0.1125, + "step": 28500 + }, + { + "epoch": 0.444718287889187, + "grad_norm": 3.6342995166778564, + "learning_rate": 0.00017035315405253637, + "loss": 0.2202, + "step": 28510 + }, + { + "epoch": 0.4448742746615087, + "grad_norm": 0.25164470076560974, + "learning_rate": 0.00017034275493438156, + "loss": 0.3322, + "step": 28520 + }, + { + "epoch": 0.4450302614338304, + "grad_norm": 3.6913859844207764, + "learning_rate": 0.0001703323558162268, + "loss": 0.3682, + "step": 28530 + }, + { + "epoch": 0.4451862482061521, + "grad_norm": 0.921101987361908, + "learning_rate": 0.000170321956698072, + "loss": 0.1553, + "step": 28540 + }, + { + "epoch": 0.4453422349784738, + "grad_norm": 0.27318769693374634, + "learning_rate": 0.00017031155757991724, + "loss": 0.1047, + "step": 28550 + }, + { + "epoch": 0.4454982217507955, + "grad_norm": 0.9199240803718567, + "learning_rate": 0.00017030115846176244, + "loss": 0.264, + "step": 28560 + }, + { + "epoch": 0.4456542085231172, + "grad_norm": 2.395423650741577, + "learning_rate": 0.00017029075934360768, + "loss": 0.2904, + "step": 28570 + }, + { + "epoch": 0.44581019529543897, + "grad_norm": 1.6989662647247314, + "learning_rate": 0.00017028036022545288, + "loss": 0.1862, + "step": 28580 + }, + { + "epoch": 0.4459661820677607, + "grad_norm": 0.7631382942199707, + "learning_rate": 0.00017026996110729812, + "loss": 0.2196, + "step": 28590 + }, + { + "epoch": 0.4461221688400824, + "grad_norm": 2.1081128120422363, + "learning_rate": 0.00017025956198914332, + "loss": 0.1705, + "step": 28600 + }, + { + "epoch": 0.4462781556124041, + "grad_norm": 1.5460604429244995, + "learning_rate": 0.00017024916287098856, + "loss": 0.1826, + "step": 28610 + }, + { + "epoch": 0.4464341423847258, + "grad_norm": 5.088230609893799, + "learning_rate": 0.00017023876375283376, + "loss": 0.2085, + "step": 28620 + }, + { + "epoch": 0.4465901291570475, + "grad_norm": 2.258603811264038, + "learning_rate": 0.000170228364634679, + "loss": 0.1396, + "step": 28630 + }, + { + "epoch": 0.4467461159293692, + "grad_norm": 1.0345207452774048, + "learning_rate": 0.0001702179655165242, + "loss": 0.3221, + "step": 28640 + }, + { + "epoch": 0.4469021027016909, + "grad_norm": 0.01587017811834812, + "learning_rate": 0.00017020756639836944, + "loss": 0.2255, + "step": 28650 + }, + { + "epoch": 0.4470580894740126, + "grad_norm": 0.4622754156589508, + "learning_rate": 0.00017019716728021463, + "loss": 0.1564, + "step": 28660 + }, + { + "epoch": 0.4472140762463343, + "grad_norm": 0.1939149796962738, + "learning_rate": 0.00017018676816205988, + "loss": 0.2337, + "step": 28670 + }, + { + "epoch": 0.447370063018656, + "grad_norm": 0.06893238425254822, + "learning_rate": 0.00017017636904390507, + "loss": 0.2173, + "step": 28680 + }, + { + "epoch": 0.4475260497909777, + "grad_norm": 1.1819541454315186, + "learning_rate": 0.00017016596992575032, + "loss": 0.2094, + "step": 28690 + }, + { + "epoch": 0.4476820365632994, + "grad_norm": 2.155634641647339, + "learning_rate": 0.0001701555708075955, + "loss": 0.4913, + "step": 28700 + }, + { + "epoch": 0.4478380233356211, + "grad_norm": 1.0332874059677124, + "learning_rate": 0.00017014517168944076, + "loss": 0.2635, + "step": 28710 + }, + { + "epoch": 0.4479940101079428, + "grad_norm": 0.33677852153778076, + "learning_rate": 0.00017013477257128595, + "loss": 0.4719, + "step": 28720 + }, + { + "epoch": 0.4481499968802646, + "grad_norm": 0.43131670355796814, + "learning_rate": 0.0001701243734531312, + "loss": 0.4657, + "step": 28730 + }, + { + "epoch": 0.4483059836525863, + "grad_norm": 2.705571413040161, + "learning_rate": 0.0001701139743349764, + "loss": 0.263, + "step": 28740 + }, + { + "epoch": 0.448461970424908, + "grad_norm": 1.7403215169906616, + "learning_rate": 0.00017010357521682164, + "loss": 0.2873, + "step": 28750 + }, + { + "epoch": 0.4486179571972297, + "grad_norm": 0.184366837143898, + "learning_rate": 0.00017009317609866683, + "loss": 0.1444, + "step": 28760 + }, + { + "epoch": 0.4487739439695514, + "grad_norm": 0.26151204109191895, + "learning_rate": 0.00017008277698051208, + "loss": 0.1371, + "step": 28770 + }, + { + "epoch": 0.4489299307418731, + "grad_norm": 0.0878930315375328, + "learning_rate": 0.00017007237786235727, + "loss": 0.2372, + "step": 28780 + }, + { + "epoch": 0.4490859175141948, + "grad_norm": 1.9251577854156494, + "learning_rate": 0.00017006197874420252, + "loss": 0.1818, + "step": 28790 + }, + { + "epoch": 0.4492419042865165, + "grad_norm": 2.9742820262908936, + "learning_rate": 0.0001700515796260477, + "loss": 0.3391, + "step": 28800 + }, + { + "epoch": 0.4493978910588382, + "grad_norm": 3.1202166080474854, + "learning_rate": 0.00017004118050789296, + "loss": 0.2956, + "step": 28810 + }, + { + "epoch": 0.4495538778311599, + "grad_norm": 0.6640981435775757, + "learning_rate": 0.00017003078138973815, + "loss": 0.2313, + "step": 28820 + }, + { + "epoch": 0.4497098646034816, + "grad_norm": 0.7795690894126892, + "learning_rate": 0.00017002038227158337, + "loss": 0.1291, + "step": 28830 + }, + { + "epoch": 0.4498658513758033, + "grad_norm": 1.0500304698944092, + "learning_rate": 0.0001700099831534286, + "loss": 0.1449, + "step": 28840 + }, + { + "epoch": 0.450021838148125, + "grad_norm": 0.5551002025604248, + "learning_rate": 0.0001699995840352738, + "loss": 0.2229, + "step": 28850 + }, + { + "epoch": 0.4501778249204467, + "grad_norm": 0.5139710307121277, + "learning_rate": 0.00016998918491711903, + "loss": 0.0654, + "step": 28860 + }, + { + "epoch": 0.4503338116927684, + "grad_norm": 2.027582883834839, + "learning_rate": 0.00016997878579896425, + "loss": 0.2082, + "step": 28870 + }, + { + "epoch": 0.4504897984650902, + "grad_norm": 3.4057729244232178, + "learning_rate": 0.00016996838668080947, + "loss": 0.1878, + "step": 28880 + }, + { + "epoch": 0.4506457852374119, + "grad_norm": 0.4414885342121124, + "learning_rate": 0.00016995798756265469, + "loss": 0.2055, + "step": 28890 + }, + { + "epoch": 0.4508017720097336, + "grad_norm": 5.074105739593506, + "learning_rate": 0.0001699475884444999, + "loss": 0.2078, + "step": 28900 + }, + { + "epoch": 0.4509577587820553, + "grad_norm": 0.23571071028709412, + "learning_rate": 0.00016993718932634512, + "loss": 0.2825, + "step": 28910 + }, + { + "epoch": 0.451113745554377, + "grad_norm": 1.4560014009475708, + "learning_rate": 0.00016992679020819034, + "loss": 0.1286, + "step": 28920 + }, + { + "epoch": 0.4512697323266987, + "grad_norm": 3.137782335281372, + "learning_rate": 0.00016991639109003556, + "loss": 0.4612, + "step": 28930 + }, + { + "epoch": 0.4514257190990204, + "grad_norm": 0.40585848689079285, + "learning_rate": 0.00016990599197188078, + "loss": 0.1337, + "step": 28940 + }, + { + "epoch": 0.4515817058713421, + "grad_norm": 0.9361881017684937, + "learning_rate": 0.000169895592853726, + "loss": 0.3224, + "step": 28950 + }, + { + "epoch": 0.4517376926436638, + "grad_norm": 0.6384800672531128, + "learning_rate": 0.00016988519373557122, + "loss": 0.3686, + "step": 28960 + }, + { + "epoch": 0.4518936794159855, + "grad_norm": 2.370234251022339, + "learning_rate": 0.00016987479461741644, + "loss": 0.223, + "step": 28970 + }, + { + "epoch": 0.4520496661883072, + "grad_norm": 1.9556089639663696, + "learning_rate": 0.00016986439549926166, + "loss": 0.182, + "step": 28980 + }, + { + "epoch": 0.4522056529606289, + "grad_norm": 0.9906886219978333, + "learning_rate": 0.00016985399638110688, + "loss": 0.0844, + "step": 28990 + }, + { + "epoch": 0.4523616397329506, + "grad_norm": 0.8223429322242737, + "learning_rate": 0.0001698435972629521, + "loss": 0.2251, + "step": 29000 + }, + { + "epoch": 0.45251762650527233, + "grad_norm": 0.08095703274011612, + "learning_rate": 0.00016983319814479732, + "loss": 0.2306, + "step": 29010 + }, + { + "epoch": 0.45267361327759403, + "grad_norm": 1.9347952604293823, + "learning_rate": 0.00016982279902664254, + "loss": 0.3436, + "step": 29020 + }, + { + "epoch": 0.4528296000499158, + "grad_norm": 0.4428502321243286, + "learning_rate": 0.00016981239990848776, + "loss": 0.2556, + "step": 29030 + }, + { + "epoch": 0.4529855868222375, + "grad_norm": 0.9700573682785034, + "learning_rate": 0.00016980200079033298, + "loss": 0.1827, + "step": 29040 + }, + { + "epoch": 0.4531415735945592, + "grad_norm": 2.3607490062713623, + "learning_rate": 0.0001697916016721782, + "loss": 0.1541, + "step": 29050 + }, + { + "epoch": 0.4532975603668809, + "grad_norm": 1.0256181955337524, + "learning_rate": 0.00016978120255402345, + "loss": 0.343, + "step": 29060 + }, + { + "epoch": 0.4534535471392026, + "grad_norm": 1.3614492416381836, + "learning_rate": 0.00016977080343586864, + "loss": 0.1686, + "step": 29070 + }, + { + "epoch": 0.4536095339115243, + "grad_norm": 3.9801783561706543, + "learning_rate": 0.00016976040431771389, + "loss": 0.2275, + "step": 29080 + }, + { + "epoch": 0.453765520683846, + "grad_norm": 2.9337174892425537, + "learning_rate": 0.00016975000519955908, + "loss": 0.2286, + "step": 29090 + }, + { + "epoch": 0.4539215074561677, + "grad_norm": 2.2115631103515625, + "learning_rate": 0.00016973960608140433, + "loss": 0.2058, + "step": 29100 + }, + { + "epoch": 0.4540774942284894, + "grad_norm": 1.9049028158187866, + "learning_rate": 0.00016972920696324952, + "loss": 0.2951, + "step": 29110 + }, + { + "epoch": 0.4542334810008111, + "grad_norm": 1.4368748664855957, + "learning_rate": 0.00016971880784509476, + "loss": 0.1578, + "step": 29120 + }, + { + "epoch": 0.4543894677731328, + "grad_norm": 1.129590392112732, + "learning_rate": 0.00016970840872693996, + "loss": 0.1412, + "step": 29130 + }, + { + "epoch": 0.45454545454545453, + "grad_norm": 8.13650131225586, + "learning_rate": 0.0001696980096087852, + "loss": 0.226, + "step": 29140 + }, + { + "epoch": 0.45470144131777623, + "grad_norm": 2.122213125228882, + "learning_rate": 0.0001696876104906304, + "loss": 0.2045, + "step": 29150 + }, + { + "epoch": 0.45485742809009794, + "grad_norm": 0.25566190481185913, + "learning_rate": 0.00016967721137247564, + "loss": 0.1355, + "step": 29160 + }, + { + "epoch": 0.45501341486241964, + "grad_norm": 1.085450530052185, + "learning_rate": 0.00016966681225432084, + "loss": 0.1669, + "step": 29170 + }, + { + "epoch": 0.4551694016347414, + "grad_norm": 1.4198920726776123, + "learning_rate": 0.00016965641313616608, + "loss": 0.127, + "step": 29180 + }, + { + "epoch": 0.4553253884070631, + "grad_norm": 0.9510000944137573, + "learning_rate": 0.00016964601401801127, + "loss": 0.1677, + "step": 29190 + }, + { + "epoch": 0.4554813751793848, + "grad_norm": 0.1792433261871338, + "learning_rate": 0.00016963561489985652, + "loss": 0.2053, + "step": 29200 + }, + { + "epoch": 0.4556373619517065, + "grad_norm": 1.2783483266830444, + "learning_rate": 0.00016962521578170171, + "loss": 0.3974, + "step": 29210 + }, + { + "epoch": 0.4557933487240282, + "grad_norm": 0.6709703803062439, + "learning_rate": 0.00016961481666354696, + "loss": 0.1778, + "step": 29220 + }, + { + "epoch": 0.4559493354963499, + "grad_norm": 2.382859230041504, + "learning_rate": 0.00016960441754539215, + "loss": 0.3592, + "step": 29230 + }, + { + "epoch": 0.4561053222686716, + "grad_norm": 0.11763356626033783, + "learning_rate": 0.0001695940184272374, + "loss": 0.1797, + "step": 29240 + }, + { + "epoch": 0.4562613090409933, + "grad_norm": 3.912564754486084, + "learning_rate": 0.0001695836193090826, + "loss": 0.4059, + "step": 29250 + }, + { + "epoch": 0.456417295813315, + "grad_norm": 0.3952312171459198, + "learning_rate": 0.00016957322019092784, + "loss": 0.1293, + "step": 29260 + }, + { + "epoch": 0.45657328258563673, + "grad_norm": 1.483230471611023, + "learning_rate": 0.00016956282107277303, + "loss": 0.1765, + "step": 29270 + }, + { + "epoch": 0.45672926935795843, + "grad_norm": 1.6304653882980347, + "learning_rate": 0.00016955242195461825, + "loss": 0.075, + "step": 29280 + }, + { + "epoch": 0.45688525613028014, + "grad_norm": 0.960439145565033, + "learning_rate": 0.00016954202283646347, + "loss": 0.133, + "step": 29290 + }, + { + "epoch": 0.45704124290260184, + "grad_norm": 1.7323864698410034, + "learning_rate": 0.0001695316237183087, + "loss": 0.1739, + "step": 29300 + }, + { + "epoch": 0.45719722967492354, + "grad_norm": 0.01141402218490839, + "learning_rate": 0.0001695212246001539, + "loss": 0.248, + "step": 29310 + }, + { + "epoch": 0.45735321644724525, + "grad_norm": 4.076611518859863, + "learning_rate": 0.00016951082548199913, + "loss": 0.387, + "step": 29320 + }, + { + "epoch": 0.457509203219567, + "grad_norm": 0.18890061974525452, + "learning_rate": 0.00016950042636384435, + "loss": 0.3097, + "step": 29330 + }, + { + "epoch": 0.4576651899918887, + "grad_norm": 0.2957797944545746, + "learning_rate": 0.00016949002724568957, + "loss": 0.2065, + "step": 29340 + }, + { + "epoch": 0.4578211767642104, + "grad_norm": 2.4584219455718994, + "learning_rate": 0.0001694796281275348, + "loss": 0.2741, + "step": 29350 + }, + { + "epoch": 0.4579771635365321, + "grad_norm": 2.8883042335510254, + "learning_rate": 0.00016946922900938, + "loss": 0.4521, + "step": 29360 + }, + { + "epoch": 0.4581331503088538, + "grad_norm": 0.7281550765037537, + "learning_rate": 0.00016945882989122523, + "loss": 0.2216, + "step": 29370 + }, + { + "epoch": 0.4582891370811755, + "grad_norm": 0.1314631849527359, + "learning_rate": 0.00016944843077307045, + "loss": 0.3388, + "step": 29380 + }, + { + "epoch": 0.4584451238534972, + "grad_norm": 0.1755414754152298, + "learning_rate": 0.00016943803165491567, + "loss": 0.2143, + "step": 29390 + }, + { + "epoch": 0.45860111062581893, + "grad_norm": 0.4767131209373474, + "learning_rate": 0.0001694276325367609, + "loss": 0.2321, + "step": 29400 + }, + { + "epoch": 0.45875709739814063, + "grad_norm": 0.43903854489326477, + "learning_rate": 0.0001694172334186061, + "loss": 0.1635, + "step": 29410 + }, + { + "epoch": 0.45891308417046234, + "grad_norm": 1.1687424182891846, + "learning_rate": 0.00016940683430045133, + "loss": 0.1434, + "step": 29420 + }, + { + "epoch": 0.45906907094278404, + "grad_norm": 2.150815725326538, + "learning_rate": 0.00016939643518229655, + "loss": 0.4474, + "step": 29430 + }, + { + "epoch": 0.45922505771510574, + "grad_norm": 1.354557752609253, + "learning_rate": 0.00016938603606414177, + "loss": 0.1259, + "step": 29440 + }, + { + "epoch": 0.45938104448742745, + "grad_norm": 1.0940755605697632, + "learning_rate": 0.00016937563694598699, + "loss": 0.2751, + "step": 29450 + }, + { + "epoch": 0.45953703125974915, + "grad_norm": 1.485103964805603, + "learning_rate": 0.0001693652378278322, + "loss": 0.383, + "step": 29460 + }, + { + "epoch": 0.45969301803207085, + "grad_norm": 0.2681877017021179, + "learning_rate": 0.00016935483870967742, + "loss": 0.1578, + "step": 29470 + }, + { + "epoch": 0.4598490048043926, + "grad_norm": 1.7786865234375, + "learning_rate": 0.00016934443959152264, + "loss": 0.1364, + "step": 29480 + }, + { + "epoch": 0.4600049915767143, + "grad_norm": 2.6613898277282715, + "learning_rate": 0.00016933404047336786, + "loss": 0.2088, + "step": 29490 + }, + { + "epoch": 0.460160978349036, + "grad_norm": 1.7964385747909546, + "learning_rate": 0.00016932364135521308, + "loss": 0.2973, + "step": 29500 + }, + { + "epoch": 0.4603169651213577, + "grad_norm": 0.7146076560020447, + "learning_rate": 0.0001693132422370583, + "loss": 0.4392, + "step": 29510 + }, + { + "epoch": 0.4604729518936794, + "grad_norm": 1.6241596937179565, + "learning_rate": 0.00016930284311890352, + "loss": 0.3902, + "step": 29520 + }, + { + "epoch": 0.46062893866600113, + "grad_norm": 0.15453019738197327, + "learning_rate": 0.00016929244400074874, + "loss": 0.2751, + "step": 29530 + }, + { + "epoch": 0.46078492543832283, + "grad_norm": 0.6671075820922852, + "learning_rate": 0.00016928204488259396, + "loss": 0.1299, + "step": 29540 + }, + { + "epoch": 0.46094091221064454, + "grad_norm": 0.06514684110879898, + "learning_rate": 0.00016927164576443918, + "loss": 0.1766, + "step": 29550 + }, + { + "epoch": 0.46109689898296624, + "grad_norm": 0.6658110022544861, + "learning_rate": 0.0001692612466462844, + "loss": 0.287, + "step": 29560 + }, + { + "epoch": 0.46125288575528794, + "grad_norm": 2.1909286975860596, + "learning_rate": 0.00016925084752812962, + "loss": 0.3568, + "step": 29570 + }, + { + "epoch": 0.46140887252760965, + "grad_norm": 0.6156850457191467, + "learning_rate": 0.00016924044840997484, + "loss": 0.165, + "step": 29580 + }, + { + "epoch": 0.46156485929993135, + "grad_norm": 1.2572039365768433, + "learning_rate": 0.00016923004929182006, + "loss": 0.1324, + "step": 29590 + }, + { + "epoch": 0.46172084607225305, + "grad_norm": 2.650160789489746, + "learning_rate": 0.00016921965017366528, + "loss": 0.2397, + "step": 29600 + }, + { + "epoch": 0.46187683284457476, + "grad_norm": 2.028608798980713, + "learning_rate": 0.0001692092510555105, + "loss": 0.1004, + "step": 29610 + }, + { + "epoch": 0.46203281961689646, + "grad_norm": 1.8734807968139648, + "learning_rate": 0.00016919885193735572, + "loss": 0.2855, + "step": 29620 + }, + { + "epoch": 0.4621888063892182, + "grad_norm": 3.063603639602661, + "learning_rate": 0.00016918845281920094, + "loss": 0.382, + "step": 29630 + }, + { + "epoch": 0.4623447931615399, + "grad_norm": 0.16561317443847656, + "learning_rate": 0.00016917805370104616, + "loss": 0.2907, + "step": 29640 + }, + { + "epoch": 0.4625007799338616, + "grad_norm": 2.0679848194122314, + "learning_rate": 0.00016916765458289138, + "loss": 0.2754, + "step": 29650 + }, + { + "epoch": 0.46265676670618333, + "grad_norm": 0.4315131902694702, + "learning_rate": 0.0001691572554647366, + "loss": 0.1671, + "step": 29660 + }, + { + "epoch": 0.46281275347850503, + "grad_norm": 0.22854287922382355, + "learning_rate": 0.00016914685634658182, + "loss": 0.1405, + "step": 29670 + }, + { + "epoch": 0.46296874025082674, + "grad_norm": 0.5000004768371582, + "learning_rate": 0.00016913645722842704, + "loss": 0.1788, + "step": 29680 + }, + { + "epoch": 0.46312472702314844, + "grad_norm": 1.142099380493164, + "learning_rate": 0.00016912605811027226, + "loss": 0.1553, + "step": 29690 + }, + { + "epoch": 0.46328071379547014, + "grad_norm": 1.4891505241394043, + "learning_rate": 0.00016911565899211748, + "loss": 0.0908, + "step": 29700 + }, + { + "epoch": 0.46343670056779185, + "grad_norm": 2.679250717163086, + "learning_rate": 0.0001691052598739627, + "loss": 0.4153, + "step": 29710 + }, + { + "epoch": 0.46359268734011355, + "grad_norm": 5.326397895812988, + "learning_rate": 0.00016909486075580792, + "loss": 0.1157, + "step": 29720 + }, + { + "epoch": 0.46374867411243526, + "grad_norm": 2.6339733600616455, + "learning_rate": 0.00016908446163765314, + "loss": 0.307, + "step": 29730 + }, + { + "epoch": 0.46390466088475696, + "grad_norm": 0.29082900285720825, + "learning_rate": 0.00016907406251949835, + "loss": 0.122, + "step": 29740 + }, + { + "epoch": 0.46406064765707866, + "grad_norm": 0.693114697933197, + "learning_rate": 0.00016906366340134357, + "loss": 0.3094, + "step": 29750 + }, + { + "epoch": 0.46421663442940037, + "grad_norm": 1.5531766414642334, + "learning_rate": 0.0001690532642831888, + "loss": 0.1714, + "step": 29760 + }, + { + "epoch": 0.46437262120172207, + "grad_norm": 1.0623689889907837, + "learning_rate": 0.00016904286516503401, + "loss": 0.1939, + "step": 29770 + }, + { + "epoch": 0.46452860797404383, + "grad_norm": 2.1284306049346924, + "learning_rate": 0.00016903246604687923, + "loss": 0.2154, + "step": 29780 + }, + { + "epoch": 0.46468459474636553, + "grad_norm": 3.249941110610962, + "learning_rate": 0.00016902206692872445, + "loss": 0.1657, + "step": 29790 + }, + { + "epoch": 0.46484058151868723, + "grad_norm": 0.9895259141921997, + "learning_rate": 0.00016901166781056967, + "loss": 0.2913, + "step": 29800 + }, + { + "epoch": 0.46499656829100894, + "grad_norm": 1.4847838878631592, + "learning_rate": 0.0001690012686924149, + "loss": 0.333, + "step": 29810 + }, + { + "epoch": 0.46515255506333064, + "grad_norm": 2.0533788204193115, + "learning_rate": 0.0001689908695742601, + "loss": 0.1995, + "step": 29820 + }, + { + "epoch": 0.46530854183565235, + "grad_norm": 1.4027974605560303, + "learning_rate": 0.00016898047045610533, + "loss": 0.2666, + "step": 29830 + }, + { + "epoch": 0.46546452860797405, + "grad_norm": 2.2192487716674805, + "learning_rate": 0.00016897007133795055, + "loss": 0.2017, + "step": 29840 + }, + { + "epoch": 0.46562051538029575, + "grad_norm": 1.4089699983596802, + "learning_rate": 0.00016895967221979577, + "loss": 0.188, + "step": 29850 + }, + { + "epoch": 0.46577650215261746, + "grad_norm": 0.45173120498657227, + "learning_rate": 0.000168949273101641, + "loss": 0.1757, + "step": 29860 + }, + { + "epoch": 0.46593248892493916, + "grad_norm": 0.9926403164863586, + "learning_rate": 0.0001689388739834862, + "loss": 0.3053, + "step": 29870 + }, + { + "epoch": 0.46608847569726086, + "grad_norm": 0.021788373589515686, + "learning_rate": 0.00016892847486533143, + "loss": 0.2482, + "step": 29880 + }, + { + "epoch": 0.46624446246958257, + "grad_norm": 1.0584843158721924, + "learning_rate": 0.00016891807574717665, + "loss": 0.2482, + "step": 29890 + }, + { + "epoch": 0.46640044924190427, + "grad_norm": 1.7717550992965698, + "learning_rate": 0.00016890767662902187, + "loss": 0.3688, + "step": 29900 + }, + { + "epoch": 0.466556436014226, + "grad_norm": 1.7900004386901855, + "learning_rate": 0.0001688972775108671, + "loss": 0.0789, + "step": 29910 + }, + { + "epoch": 0.4667124227865477, + "grad_norm": 2.7771332263946533, + "learning_rate": 0.0001688868783927123, + "loss": 0.2271, + "step": 29920 + }, + { + "epoch": 0.46686840955886943, + "grad_norm": 2.65687894821167, + "learning_rate": 0.00016887647927455753, + "loss": 0.5232, + "step": 29930 + }, + { + "epoch": 0.46702439633119114, + "grad_norm": 0.2827414572238922, + "learning_rate": 0.00016886608015640275, + "loss": 0.2636, + "step": 29940 + }, + { + "epoch": 0.46718038310351284, + "grad_norm": 0.3519066572189331, + "learning_rate": 0.00016885568103824797, + "loss": 0.1225, + "step": 29950 + }, + { + "epoch": 0.46733636987583455, + "grad_norm": 0.3117619454860687, + "learning_rate": 0.0001688452819200932, + "loss": 0.2189, + "step": 29960 + }, + { + "epoch": 0.46749235664815625, + "grad_norm": 3.3366892337799072, + "learning_rate": 0.0001688348828019384, + "loss": 0.0941, + "step": 29970 + }, + { + "epoch": 0.46764834342047795, + "grad_norm": 0.4896647036075592, + "learning_rate": 0.00016882448368378363, + "loss": 0.264, + "step": 29980 + }, + { + "epoch": 0.46780433019279966, + "grad_norm": 0.6250243186950684, + "learning_rate": 0.00016881408456562885, + "loss": 0.281, + "step": 29990 + }, + { + "epoch": 0.46796031696512136, + "grad_norm": 2.8796308040618896, + "learning_rate": 0.00016880368544747407, + "loss": 0.1444, + "step": 30000 + }, + { + "epoch": 0.46811630373744306, + "grad_norm": 0.45706239342689514, + "learning_rate": 0.00016879328632931929, + "loss": 0.2803, + "step": 30010 + }, + { + "epoch": 0.46827229050976477, + "grad_norm": 3.09799861907959, + "learning_rate": 0.0001687828872111645, + "loss": 0.4451, + "step": 30020 + }, + { + "epoch": 0.46842827728208647, + "grad_norm": 1.12473726272583, + "learning_rate": 0.00016877248809300972, + "loss": 0.3071, + "step": 30030 + }, + { + "epoch": 0.4685842640544082, + "grad_norm": 0.03669362887740135, + "learning_rate": 0.00016876208897485494, + "loss": 0.1993, + "step": 30040 + }, + { + "epoch": 0.4687402508267299, + "grad_norm": 0.9411380290985107, + "learning_rate": 0.00016875168985670016, + "loss": 0.2847, + "step": 30050 + }, + { + "epoch": 0.4688962375990516, + "grad_norm": 1.3741745948791504, + "learning_rate": 0.00016874129073854538, + "loss": 0.233, + "step": 30060 + }, + { + "epoch": 0.4690522243713733, + "grad_norm": 1.0064387321472168, + "learning_rate": 0.0001687308916203906, + "loss": 0.1907, + "step": 30070 + }, + { + "epoch": 0.46920821114369504, + "grad_norm": 0.5974047183990479, + "learning_rate": 0.00016872049250223582, + "loss": 0.2311, + "step": 30080 + }, + { + "epoch": 0.46936419791601675, + "grad_norm": 1.5840703248977661, + "learning_rate": 0.00016871009338408104, + "loss": 0.1995, + "step": 30090 + }, + { + "epoch": 0.46952018468833845, + "grad_norm": 2.8465263843536377, + "learning_rate": 0.00016869969426592626, + "loss": 0.1383, + "step": 30100 + }, + { + "epoch": 0.46967617146066015, + "grad_norm": 3.1921279430389404, + "learning_rate": 0.00016868929514777148, + "loss": 0.241, + "step": 30110 + }, + { + "epoch": 0.46983215823298186, + "grad_norm": 2.8651912212371826, + "learning_rate": 0.0001686788960296167, + "loss": 0.5101, + "step": 30120 + }, + { + "epoch": 0.46998814500530356, + "grad_norm": 0.3068169951438904, + "learning_rate": 0.00016866849691146192, + "loss": 0.2368, + "step": 30130 + }, + { + "epoch": 0.47014413177762526, + "grad_norm": 2.445636749267578, + "learning_rate": 0.00016865809779330714, + "loss": 0.3581, + "step": 30140 + }, + { + "epoch": 0.47030011854994697, + "grad_norm": 2.4613261222839355, + "learning_rate": 0.00016864769867515236, + "loss": 0.1873, + "step": 30150 + }, + { + "epoch": 0.47045610532226867, + "grad_norm": 1.1140657663345337, + "learning_rate": 0.00016863729955699758, + "loss": 0.205, + "step": 30160 + }, + { + "epoch": 0.4706120920945904, + "grad_norm": 1.094092607498169, + "learning_rate": 0.0001686269004388428, + "loss": 0.1597, + "step": 30170 + }, + { + "epoch": 0.4707680788669121, + "grad_norm": 2.7741963863372803, + "learning_rate": 0.000168616501320688, + "loss": 0.1787, + "step": 30180 + }, + { + "epoch": 0.4709240656392338, + "grad_norm": 0.01928747072815895, + "learning_rate": 0.00016860610220253324, + "loss": 0.3335, + "step": 30190 + }, + { + "epoch": 0.4710800524115555, + "grad_norm": 6.315832614898682, + "learning_rate": 0.00016859570308437843, + "loss": 0.1956, + "step": 30200 + }, + { + "epoch": 0.4712360391838772, + "grad_norm": 2.0129244327545166, + "learning_rate": 0.00016858530396622368, + "loss": 0.2482, + "step": 30210 + }, + { + "epoch": 0.4713920259561989, + "grad_norm": 0.9225463271141052, + "learning_rate": 0.00016857490484806887, + "loss": 0.2767, + "step": 30220 + }, + { + "epoch": 0.47154801272852065, + "grad_norm": 0.07704778760671616, + "learning_rate": 0.00016856450572991412, + "loss": 0.086, + "step": 30230 + }, + { + "epoch": 0.47170399950084235, + "grad_norm": 1.3533331155776978, + "learning_rate": 0.0001685541066117593, + "loss": 0.3803, + "step": 30240 + }, + { + "epoch": 0.47185998627316406, + "grad_norm": 1.5062912702560425, + "learning_rate": 0.00016854370749360456, + "loss": 0.4857, + "step": 30250 + }, + { + "epoch": 0.47201597304548576, + "grad_norm": 3.7081758975982666, + "learning_rate": 0.00016853330837544975, + "loss": 0.4355, + "step": 30260 + }, + { + "epoch": 0.47217195981780746, + "grad_norm": 0.6643535494804382, + "learning_rate": 0.000168522909257295, + "loss": 0.1789, + "step": 30270 + }, + { + "epoch": 0.47232794659012917, + "grad_norm": 1.025472640991211, + "learning_rate": 0.0001685125101391402, + "loss": 0.3245, + "step": 30280 + }, + { + "epoch": 0.47248393336245087, + "grad_norm": 0.02963162399828434, + "learning_rate": 0.00016850211102098544, + "loss": 0.2653, + "step": 30290 + }, + { + "epoch": 0.4726399201347726, + "grad_norm": 0.904206395149231, + "learning_rate": 0.00016849171190283063, + "loss": 0.3088, + "step": 30300 + }, + { + "epoch": 0.4727959069070943, + "grad_norm": 0.03180241584777832, + "learning_rate": 0.00016848131278467587, + "loss": 0.2732, + "step": 30310 + }, + { + "epoch": 0.472951893679416, + "grad_norm": 3.9663686752319336, + "learning_rate": 0.00016847091366652107, + "loss": 0.2548, + "step": 30320 + }, + { + "epoch": 0.4731078804517377, + "grad_norm": 1.9202165603637695, + "learning_rate": 0.00016846051454836631, + "loss": 0.1617, + "step": 30330 + }, + { + "epoch": 0.4732638672240594, + "grad_norm": 0.32910484075546265, + "learning_rate": 0.0001684501154302115, + "loss": 0.3312, + "step": 30340 + }, + { + "epoch": 0.4734198539963811, + "grad_norm": 0.2507750988006592, + "learning_rate": 0.00016843971631205675, + "loss": 0.1519, + "step": 30350 + }, + { + "epoch": 0.4735758407687028, + "grad_norm": 2.988393783569336, + "learning_rate": 0.00016842931719390195, + "loss": 0.2504, + "step": 30360 + }, + { + "epoch": 0.4737318275410245, + "grad_norm": 1.0900479555130005, + "learning_rate": 0.0001684189180757472, + "loss": 0.1603, + "step": 30370 + }, + { + "epoch": 0.4738878143133462, + "grad_norm": 1.146888256072998, + "learning_rate": 0.00016840851895759238, + "loss": 0.1461, + "step": 30380 + }, + { + "epoch": 0.47404380108566796, + "grad_norm": 2.4643445014953613, + "learning_rate": 0.00016839811983943763, + "loss": 0.2443, + "step": 30390 + }, + { + "epoch": 0.47419978785798966, + "grad_norm": 1.1018608808517456, + "learning_rate": 0.00016838772072128285, + "loss": 0.1733, + "step": 30400 + }, + { + "epoch": 0.47435577463031137, + "grad_norm": 0.9486880898475647, + "learning_rate": 0.00016837732160312807, + "loss": 0.2001, + "step": 30410 + }, + { + "epoch": 0.47451176140263307, + "grad_norm": 2.11059308052063, + "learning_rate": 0.0001683669224849733, + "loss": 0.3697, + "step": 30420 + }, + { + "epoch": 0.4746677481749548, + "grad_norm": 2.2037315368652344, + "learning_rate": 0.0001683565233668185, + "loss": 0.1812, + "step": 30430 + }, + { + "epoch": 0.4748237349472765, + "grad_norm": 4.476726055145264, + "learning_rate": 0.00016834612424866373, + "loss": 0.2603, + "step": 30440 + }, + { + "epoch": 0.4749797217195982, + "grad_norm": 1.6905113458633423, + "learning_rate": 0.00016833572513050895, + "loss": 0.3342, + "step": 30450 + }, + { + "epoch": 0.4751357084919199, + "grad_norm": 1.99224054813385, + "learning_rate": 0.00016832532601235417, + "loss": 0.3406, + "step": 30460 + }, + { + "epoch": 0.4752916952642416, + "grad_norm": 0.21781033277511597, + "learning_rate": 0.0001683149268941994, + "loss": 0.3403, + "step": 30470 + }, + { + "epoch": 0.4754476820365633, + "grad_norm": 3.0926010608673096, + "learning_rate": 0.0001683045277760446, + "loss": 0.2239, + "step": 30480 + }, + { + "epoch": 0.475603668808885, + "grad_norm": 0.7872567772865295, + "learning_rate": 0.00016829412865788983, + "loss": 0.254, + "step": 30490 + }, + { + "epoch": 0.4757596555812067, + "grad_norm": 0.04311418533325195, + "learning_rate": 0.00016828372953973505, + "loss": 0.1023, + "step": 30500 + }, + { + "epoch": 0.4759156423535284, + "grad_norm": 2.7619473934173584, + "learning_rate": 0.00016827333042158027, + "loss": 0.2767, + "step": 30510 + }, + { + "epoch": 0.4760716291258501, + "grad_norm": 3.2754528522491455, + "learning_rate": 0.0001682629313034255, + "loss": 0.2568, + "step": 30520 + }, + { + "epoch": 0.4762276158981718, + "grad_norm": 5.349419116973877, + "learning_rate": 0.0001682525321852707, + "loss": 0.2549, + "step": 30530 + }, + { + "epoch": 0.47638360267049357, + "grad_norm": 2.556077003479004, + "learning_rate": 0.00016824213306711593, + "loss": 0.3167, + "step": 30540 + }, + { + "epoch": 0.47653958944281527, + "grad_norm": 0.722279965877533, + "learning_rate": 0.00016823173394896115, + "loss": 0.2222, + "step": 30550 + }, + { + "epoch": 0.476695576215137, + "grad_norm": 0.7086097002029419, + "learning_rate": 0.00016822133483080637, + "loss": 0.1719, + "step": 30560 + }, + { + "epoch": 0.4768515629874587, + "grad_norm": 1.0537711381912231, + "learning_rate": 0.00016821093571265159, + "loss": 0.1872, + "step": 30570 + }, + { + "epoch": 0.4770075497597804, + "grad_norm": 0.6848838329315186, + "learning_rate": 0.0001682005365944968, + "loss": 0.3955, + "step": 30580 + }, + { + "epoch": 0.4771635365321021, + "grad_norm": 1.4214539527893066, + "learning_rate": 0.00016819013747634202, + "loss": 0.3603, + "step": 30590 + }, + { + "epoch": 0.4773195233044238, + "grad_norm": 0.6845849752426147, + "learning_rate": 0.00016817973835818724, + "loss": 0.3758, + "step": 30600 + }, + { + "epoch": 0.4774755100767455, + "grad_norm": 1.2099288702011108, + "learning_rate": 0.00016816933924003246, + "loss": 0.0931, + "step": 30610 + }, + { + "epoch": 0.4776314968490672, + "grad_norm": 1.2241394519805908, + "learning_rate": 0.00016815894012187768, + "loss": 0.2159, + "step": 30620 + }, + { + "epoch": 0.4777874836213889, + "grad_norm": 2.2076525688171387, + "learning_rate": 0.00016814854100372288, + "loss": 0.1766, + "step": 30630 + }, + { + "epoch": 0.4779434703937106, + "grad_norm": 0.21678611636161804, + "learning_rate": 0.00016813814188556812, + "loss": 0.1334, + "step": 30640 + }, + { + "epoch": 0.4780994571660323, + "grad_norm": 2.96474027633667, + "learning_rate": 0.00016812774276741332, + "loss": 0.1434, + "step": 30650 + }, + { + "epoch": 0.478255443938354, + "grad_norm": 0.5001376867294312, + "learning_rate": 0.00016811734364925856, + "loss": 0.1282, + "step": 30660 + }, + { + "epoch": 0.4784114307106757, + "grad_norm": 1.8183201551437378, + "learning_rate": 0.00016810694453110375, + "loss": 0.2019, + "step": 30670 + }, + { + "epoch": 0.4785674174829974, + "grad_norm": 2.0168581008911133, + "learning_rate": 0.000168096545412949, + "loss": 0.664, + "step": 30680 + }, + { + "epoch": 0.4787234042553192, + "grad_norm": 1.144456386566162, + "learning_rate": 0.0001680861462947942, + "loss": 0.2709, + "step": 30690 + }, + { + "epoch": 0.4788793910276409, + "grad_norm": 0.24063318967819214, + "learning_rate": 0.00016807574717663944, + "loss": 0.182, + "step": 30700 + }, + { + "epoch": 0.4790353777999626, + "grad_norm": 1.794215440750122, + "learning_rate": 0.00016806534805848463, + "loss": 0.3837, + "step": 30710 + }, + { + "epoch": 0.4791913645722843, + "grad_norm": 1.3007612228393555, + "learning_rate": 0.00016805494894032988, + "loss": 0.1593, + "step": 30720 + }, + { + "epoch": 0.479347351344606, + "grad_norm": 1.4692269563674927, + "learning_rate": 0.00016804454982217507, + "loss": 0.3292, + "step": 30730 + }, + { + "epoch": 0.4795033381169277, + "grad_norm": 3.906578540802002, + "learning_rate": 0.00016803415070402032, + "loss": 0.1871, + "step": 30740 + }, + { + "epoch": 0.4796593248892494, + "grad_norm": 0.9656121730804443, + "learning_rate": 0.0001680237515858655, + "loss": 0.2847, + "step": 30750 + }, + { + "epoch": 0.4798153116615711, + "grad_norm": 0.6036972999572754, + "learning_rate": 0.00016801335246771076, + "loss": 0.3497, + "step": 30760 + }, + { + "epoch": 0.4799712984338928, + "grad_norm": 0.5647649168968201, + "learning_rate": 0.00016800295334955595, + "loss": 0.106, + "step": 30770 + }, + { + "epoch": 0.4801272852062145, + "grad_norm": 1.6186362504959106, + "learning_rate": 0.0001679925542314012, + "loss": 0.4951, + "step": 30780 + }, + { + "epoch": 0.4802832719785362, + "grad_norm": 1.5997332334518433, + "learning_rate": 0.0001679821551132464, + "loss": 0.138, + "step": 30790 + }, + { + "epoch": 0.4804392587508579, + "grad_norm": 2.1405560970306396, + "learning_rate": 0.00016797175599509164, + "loss": 0.2783, + "step": 30800 + }, + { + "epoch": 0.4805952455231796, + "grad_norm": 3.2609453201293945, + "learning_rate": 0.00016796135687693683, + "loss": 0.4811, + "step": 30810 + }, + { + "epoch": 0.4807512322955013, + "grad_norm": 0.33492788672447205, + "learning_rate": 0.00016795095775878208, + "loss": 0.26, + "step": 30820 + }, + { + "epoch": 0.480907219067823, + "grad_norm": 2.26206374168396, + "learning_rate": 0.00016794055864062727, + "loss": 0.333, + "step": 30830 + }, + { + "epoch": 0.4810632058401448, + "grad_norm": 3.740520715713501, + "learning_rate": 0.00016793015952247252, + "loss": 0.3692, + "step": 30840 + }, + { + "epoch": 0.4812191926124665, + "grad_norm": 2.778208017349243, + "learning_rate": 0.0001679197604043177, + "loss": 0.2117, + "step": 30850 + }, + { + "epoch": 0.4813751793847882, + "grad_norm": 1.2541290521621704, + "learning_rate": 0.00016790936128616295, + "loss": 0.1514, + "step": 30860 + }, + { + "epoch": 0.4815311661571099, + "grad_norm": 0.20798040926456451, + "learning_rate": 0.00016789896216800815, + "loss": 0.1634, + "step": 30870 + }, + { + "epoch": 0.4816871529294316, + "grad_norm": 2.002734422683716, + "learning_rate": 0.0001678885630498534, + "loss": 0.2855, + "step": 30880 + }, + { + "epoch": 0.4818431397017533, + "grad_norm": 0.05174250528216362, + "learning_rate": 0.0001678781639316986, + "loss": 0.1199, + "step": 30890 + }, + { + "epoch": 0.481999126474075, + "grad_norm": 0.5534813404083252, + "learning_rate": 0.00016786776481354383, + "loss": 0.2266, + "step": 30900 + }, + { + "epoch": 0.4821551132463967, + "grad_norm": 4.476343631744385, + "learning_rate": 0.00016785736569538903, + "loss": 0.3127, + "step": 30910 + }, + { + "epoch": 0.4823111000187184, + "grad_norm": 0.3827973008155823, + "learning_rate": 0.00016784696657723427, + "loss": 0.2587, + "step": 30920 + }, + { + "epoch": 0.4824670867910401, + "grad_norm": 2.4817111492156982, + "learning_rate": 0.00016783656745907947, + "loss": 0.3426, + "step": 30930 + }, + { + "epoch": 0.4826230735633618, + "grad_norm": 4.081475734710693, + "learning_rate": 0.0001678261683409247, + "loss": 0.336, + "step": 30940 + }, + { + "epoch": 0.4827790603356835, + "grad_norm": 0.9034212231636047, + "learning_rate": 0.0001678157692227699, + "loss": 0.0658, + "step": 30950 + }, + { + "epoch": 0.4829350471080052, + "grad_norm": 0.40599894523620605, + "learning_rate": 0.00016780537010461515, + "loss": 0.4369, + "step": 30960 + }, + { + "epoch": 0.4830910338803269, + "grad_norm": 1.0981228351593018, + "learning_rate": 0.00016779497098646034, + "loss": 0.1863, + "step": 30970 + }, + { + "epoch": 0.48324702065264863, + "grad_norm": 1.3142019510269165, + "learning_rate": 0.0001677845718683056, + "loss": 0.3275, + "step": 30980 + }, + { + "epoch": 0.4834030074249704, + "grad_norm": 2.7194106578826904, + "learning_rate": 0.00016777417275015078, + "loss": 0.1971, + "step": 30990 + }, + { + "epoch": 0.4835589941972921, + "grad_norm": 0.5351995825767517, + "learning_rate": 0.00016776377363199603, + "loss": 0.2848, + "step": 31000 + }, + { + "epoch": 0.4837149809696138, + "grad_norm": 3.348212242126465, + "learning_rate": 0.00016775337451384122, + "loss": 0.2182, + "step": 31010 + }, + { + "epoch": 0.4838709677419355, + "grad_norm": 2.4917593002319336, + "learning_rate": 0.00016774297539568647, + "loss": 0.2174, + "step": 31020 + }, + { + "epoch": 0.4840269545142572, + "grad_norm": 0.5034515261650085, + "learning_rate": 0.00016773257627753166, + "loss": 0.2614, + "step": 31030 + }, + { + "epoch": 0.4841829412865789, + "grad_norm": 2.6194422245025635, + "learning_rate": 0.0001677221771593769, + "loss": 0.3049, + "step": 31040 + }, + { + "epoch": 0.4843389280589006, + "grad_norm": 2.3590071201324463, + "learning_rate": 0.0001677117780412221, + "loss": 0.4772, + "step": 31050 + }, + { + "epoch": 0.4844949148312223, + "grad_norm": 0.8845195174217224, + "learning_rate": 0.00016770137892306735, + "loss": 0.2311, + "step": 31060 + }, + { + "epoch": 0.484650901603544, + "grad_norm": 1.5952205657958984, + "learning_rate": 0.00016769097980491254, + "loss": 0.0978, + "step": 31070 + }, + { + "epoch": 0.4848068883758657, + "grad_norm": 2.432046890258789, + "learning_rate": 0.00016768058068675776, + "loss": 0.516, + "step": 31080 + }, + { + "epoch": 0.4849628751481874, + "grad_norm": 0.7153600454330444, + "learning_rate": 0.00016767018156860298, + "loss": 0.2313, + "step": 31090 + }, + { + "epoch": 0.4851188619205091, + "grad_norm": 1.3413581848144531, + "learning_rate": 0.0001676597824504482, + "loss": 0.2077, + "step": 31100 + }, + { + "epoch": 0.48527484869283083, + "grad_norm": 4.311595916748047, + "learning_rate": 0.00016764938333229342, + "loss": 0.1442, + "step": 31110 + }, + { + "epoch": 0.48543083546515253, + "grad_norm": 0.05879051610827446, + "learning_rate": 0.00016763898421413864, + "loss": 0.1958, + "step": 31120 + }, + { + "epoch": 0.48558682223747424, + "grad_norm": 0.5474847555160522, + "learning_rate": 0.00016762858509598386, + "loss": 0.1344, + "step": 31130 + }, + { + "epoch": 0.485742809009796, + "grad_norm": 3.3878626823425293, + "learning_rate": 0.00016761818597782908, + "loss": 0.3811, + "step": 31140 + }, + { + "epoch": 0.4858987957821177, + "grad_norm": 0.7229539752006531, + "learning_rate": 0.0001676077868596743, + "loss": 0.1547, + "step": 31150 + }, + { + "epoch": 0.4860547825544394, + "grad_norm": 7.216814994812012, + "learning_rate": 0.00016759738774151952, + "loss": 0.2275, + "step": 31160 + }, + { + "epoch": 0.4862107693267611, + "grad_norm": 1.036226511001587, + "learning_rate": 0.00016758698862336474, + "loss": 0.2536, + "step": 31170 + }, + { + "epoch": 0.4863667560990828, + "grad_norm": 0.9178919792175293, + "learning_rate": 0.00016757658950520996, + "loss": 0.3854, + "step": 31180 + }, + { + "epoch": 0.4865227428714045, + "grad_norm": 2.5948541164398193, + "learning_rate": 0.00016756619038705518, + "loss": 0.4112, + "step": 31190 + }, + { + "epoch": 0.4866787296437262, + "grad_norm": 2.4999775886535645, + "learning_rate": 0.0001675557912689004, + "loss": 0.2547, + "step": 31200 + }, + { + "epoch": 0.4868347164160479, + "grad_norm": 0.144962877035141, + "learning_rate": 0.00016754539215074562, + "loss": 0.1984, + "step": 31210 + }, + { + "epoch": 0.4869907031883696, + "grad_norm": 2.9588258266448975, + "learning_rate": 0.00016753499303259083, + "loss": 0.2644, + "step": 31220 + }, + { + "epoch": 0.4871466899606913, + "grad_norm": 2.5843300819396973, + "learning_rate": 0.00016752459391443605, + "loss": 0.1749, + "step": 31230 + }, + { + "epoch": 0.48730267673301303, + "grad_norm": 0.09571018815040588, + "learning_rate": 0.00016751419479628127, + "loss": 0.1627, + "step": 31240 + }, + { + "epoch": 0.48745866350533473, + "grad_norm": 0.40502116084098816, + "learning_rate": 0.00016750379567812652, + "loss": 0.102, + "step": 31250 + }, + { + "epoch": 0.48761465027765644, + "grad_norm": 0.6685170531272888, + "learning_rate": 0.0001674933965599717, + "loss": 0.1366, + "step": 31260 + }, + { + "epoch": 0.48777063704997814, + "grad_norm": 2.617445945739746, + "learning_rate": 0.00016748299744181696, + "loss": 0.376, + "step": 31270 + }, + { + "epoch": 0.48792662382229984, + "grad_norm": 0.2501496970653534, + "learning_rate": 0.00016747259832366215, + "loss": 0.1891, + "step": 31280 + }, + { + "epoch": 0.4880826105946216, + "grad_norm": 4.831740856170654, + "learning_rate": 0.0001674621992055074, + "loss": 0.1284, + "step": 31290 + }, + { + "epoch": 0.4882385973669433, + "grad_norm": 0.45422425866127014, + "learning_rate": 0.0001674518000873526, + "loss": 0.3766, + "step": 31300 + }, + { + "epoch": 0.488394584139265, + "grad_norm": 1.9848741292953491, + "learning_rate": 0.00016744140096919784, + "loss": 0.3361, + "step": 31310 + }, + { + "epoch": 0.4885505709115867, + "grad_norm": 4.921041011810303, + "learning_rate": 0.00016743100185104303, + "loss": 0.4202, + "step": 31320 + }, + { + "epoch": 0.4887065576839084, + "grad_norm": 1.487733006477356, + "learning_rate": 0.00016742060273288828, + "loss": 0.4586, + "step": 31330 + }, + { + "epoch": 0.4888625444562301, + "grad_norm": 0.6384997367858887, + "learning_rate": 0.00016741020361473347, + "loss": 0.0908, + "step": 31340 + }, + { + "epoch": 0.4890185312285518, + "grad_norm": 1.0614384412765503, + "learning_rate": 0.00016739980449657872, + "loss": 0.1983, + "step": 31350 + }, + { + "epoch": 0.4891745180008735, + "grad_norm": 1.2721771001815796, + "learning_rate": 0.0001673894053784239, + "loss": 0.4351, + "step": 31360 + }, + { + "epoch": 0.48933050477319523, + "grad_norm": 1.858496904373169, + "learning_rate": 0.00016737900626026916, + "loss": 0.2082, + "step": 31370 + }, + { + "epoch": 0.48948649154551693, + "grad_norm": 0.6314982771873474, + "learning_rate": 0.00016736860714211435, + "loss": 0.1628, + "step": 31380 + }, + { + "epoch": 0.48964247831783864, + "grad_norm": 0.921778678894043, + "learning_rate": 0.0001673582080239596, + "loss": 0.1501, + "step": 31390 + }, + { + "epoch": 0.48979846509016034, + "grad_norm": 0.31458309292793274, + "learning_rate": 0.0001673478089058048, + "loss": 0.3186, + "step": 31400 + }, + { + "epoch": 0.48995445186248204, + "grad_norm": 1.1328305006027222, + "learning_rate": 0.00016733740978765003, + "loss": 0.1806, + "step": 31410 + }, + { + "epoch": 0.49011043863480375, + "grad_norm": 1.2925907373428345, + "learning_rate": 0.00016732701066949523, + "loss": 0.1999, + "step": 31420 + }, + { + "epoch": 0.49026642540712545, + "grad_norm": 0.7245944738388062, + "learning_rate": 0.00016731661155134047, + "loss": 0.3196, + "step": 31430 + }, + { + "epoch": 0.4904224121794472, + "grad_norm": 1.4912376403808594, + "learning_rate": 0.00016730621243318567, + "loss": 0.2033, + "step": 31440 + }, + { + "epoch": 0.4905783989517689, + "grad_norm": 1.4982343912124634, + "learning_rate": 0.0001672958133150309, + "loss": 0.18, + "step": 31450 + }, + { + "epoch": 0.4907343857240906, + "grad_norm": 0.29712292551994324, + "learning_rate": 0.0001672854141968761, + "loss": 0.1476, + "step": 31460 + }, + { + "epoch": 0.4908903724964123, + "grad_norm": 0.12274292856454849, + "learning_rate": 0.00016727501507872135, + "loss": 0.1319, + "step": 31470 + }, + { + "epoch": 0.491046359268734, + "grad_norm": 2.4385111331939697, + "learning_rate": 0.00016726461596056655, + "loss": 0.1678, + "step": 31480 + }, + { + "epoch": 0.4912023460410557, + "grad_norm": 2.6319873332977295, + "learning_rate": 0.0001672542168424118, + "loss": 0.2618, + "step": 31490 + }, + { + "epoch": 0.49135833281337743, + "grad_norm": 0.5047391057014465, + "learning_rate": 0.00016724381772425698, + "loss": 0.1732, + "step": 31500 + }, + { + "epoch": 0.49151431958569913, + "grad_norm": 0.13262036442756653, + "learning_rate": 0.00016723341860610223, + "loss": 0.2248, + "step": 31510 + }, + { + "epoch": 0.49167030635802084, + "grad_norm": 2.251399040222168, + "learning_rate": 0.00016722301948794742, + "loss": 0.2365, + "step": 31520 + }, + { + "epoch": 0.49182629313034254, + "grad_norm": 1.2853648662567139, + "learning_rate": 0.00016721262036979264, + "loss": 0.3446, + "step": 31530 + }, + { + "epoch": 0.49198227990266424, + "grad_norm": 4.565437316894531, + "learning_rate": 0.00016720222125163786, + "loss": 0.2721, + "step": 31540 + }, + { + "epoch": 0.49213826667498595, + "grad_norm": 2.6204447746276855, + "learning_rate": 0.00016719182213348308, + "loss": 0.2801, + "step": 31550 + }, + { + "epoch": 0.49229425344730765, + "grad_norm": 0.6889423727989197, + "learning_rate": 0.0001671814230153283, + "loss": 0.0752, + "step": 31560 + }, + { + "epoch": 0.49245024021962935, + "grad_norm": 4.357355117797852, + "learning_rate": 0.00016717102389717352, + "loss": 0.1603, + "step": 31570 + }, + { + "epoch": 0.49260622699195106, + "grad_norm": 0.7470182180404663, + "learning_rate": 0.00016716062477901874, + "loss": 0.3114, + "step": 31580 + }, + { + "epoch": 0.4927622137642728, + "grad_norm": 2.0040745735168457, + "learning_rate": 0.00016715022566086396, + "loss": 0.235, + "step": 31590 + }, + { + "epoch": 0.4929182005365945, + "grad_norm": 0.06579481065273285, + "learning_rate": 0.00016713982654270918, + "loss": 0.1664, + "step": 31600 + }, + { + "epoch": 0.4930741873089162, + "grad_norm": 0.3582492172718048, + "learning_rate": 0.0001671294274245544, + "loss": 0.3467, + "step": 31610 + }, + { + "epoch": 0.4932301740812379, + "grad_norm": 1.3861957788467407, + "learning_rate": 0.00016711902830639962, + "loss": 0.2794, + "step": 31620 + }, + { + "epoch": 0.49338616085355963, + "grad_norm": 1.14900803565979, + "learning_rate": 0.00016710862918824484, + "loss": 0.2921, + "step": 31630 + }, + { + "epoch": 0.49354214762588133, + "grad_norm": 3.2658607959747314, + "learning_rate": 0.00016709823007009006, + "loss": 0.2749, + "step": 31640 + }, + { + "epoch": 0.49369813439820304, + "grad_norm": 2.808101177215576, + "learning_rate": 0.00016708783095193528, + "loss": 0.4017, + "step": 31650 + }, + { + "epoch": 0.49385412117052474, + "grad_norm": 1.4082258939743042, + "learning_rate": 0.0001670774318337805, + "loss": 0.2971, + "step": 31660 + }, + { + "epoch": 0.49401010794284644, + "grad_norm": 0.28355127573013306, + "learning_rate": 0.00016706703271562572, + "loss": 0.1553, + "step": 31670 + }, + { + "epoch": 0.49416609471516815, + "grad_norm": 1.8266841173171997, + "learning_rate": 0.00016705663359747094, + "loss": 0.2117, + "step": 31680 + }, + { + "epoch": 0.49432208148748985, + "grad_norm": 0.9614700675010681, + "learning_rate": 0.00016704623447931616, + "loss": 0.3848, + "step": 31690 + }, + { + "epoch": 0.49447806825981155, + "grad_norm": 2.498059034347534, + "learning_rate": 0.00016703583536116138, + "loss": 0.3739, + "step": 31700 + }, + { + "epoch": 0.49463405503213326, + "grad_norm": 0.9393208026885986, + "learning_rate": 0.0001670254362430066, + "loss": 0.2067, + "step": 31710 + }, + { + "epoch": 0.49479004180445496, + "grad_norm": 0.6228502988815308, + "learning_rate": 0.00016701503712485182, + "loss": 0.3049, + "step": 31720 + }, + { + "epoch": 0.49494602857677666, + "grad_norm": 0.5203268527984619, + "learning_rate": 0.00016700463800669704, + "loss": 0.1938, + "step": 31730 + }, + { + "epoch": 0.4951020153490984, + "grad_norm": 2.057316303253174, + "learning_rate": 0.00016699423888854226, + "loss": 0.2952, + "step": 31740 + }, + { + "epoch": 0.4952580021214201, + "grad_norm": 2.162051200866699, + "learning_rate": 0.00016698383977038748, + "loss": 0.1783, + "step": 31750 + }, + { + "epoch": 0.49541398889374183, + "grad_norm": 1.1329420804977417, + "learning_rate": 0.0001669734406522327, + "loss": 0.1312, + "step": 31760 + }, + { + "epoch": 0.49556997566606353, + "grad_norm": 0.19824473559856415, + "learning_rate": 0.00016696304153407791, + "loss": 0.3658, + "step": 31770 + }, + { + "epoch": 0.49572596243838524, + "grad_norm": 0.7616140246391296, + "learning_rate": 0.00016695264241592313, + "loss": 0.3839, + "step": 31780 + }, + { + "epoch": 0.49588194921070694, + "grad_norm": 0.25937578082084656, + "learning_rate": 0.00016694224329776835, + "loss": 0.276, + "step": 31790 + }, + { + "epoch": 0.49603793598302864, + "grad_norm": 1.5954430103302002, + "learning_rate": 0.00016693184417961357, + "loss": 0.2938, + "step": 31800 + }, + { + "epoch": 0.49619392275535035, + "grad_norm": 1.2336664199829102, + "learning_rate": 0.0001669214450614588, + "loss": 0.1058, + "step": 31810 + }, + { + "epoch": 0.49634990952767205, + "grad_norm": 1.1048794984817505, + "learning_rate": 0.000166911045943304, + "loss": 0.177, + "step": 31820 + }, + { + "epoch": 0.49650589629999375, + "grad_norm": 0.29061630368232727, + "learning_rate": 0.00016690064682514923, + "loss": 0.2392, + "step": 31830 + }, + { + "epoch": 0.49666188307231546, + "grad_norm": 3.3957440853118896, + "learning_rate": 0.00016689024770699445, + "loss": 0.1533, + "step": 31840 + }, + { + "epoch": 0.49681786984463716, + "grad_norm": 0.6493799686431885, + "learning_rate": 0.00016687984858883967, + "loss": 0.2778, + "step": 31850 + }, + { + "epoch": 0.49697385661695886, + "grad_norm": 0.35775133967399597, + "learning_rate": 0.0001668694494706849, + "loss": 0.2512, + "step": 31860 + }, + { + "epoch": 0.49712984338928057, + "grad_norm": 1.582170844078064, + "learning_rate": 0.0001668590503525301, + "loss": 0.2456, + "step": 31870 + }, + { + "epoch": 0.49728583016160227, + "grad_norm": 0.9892213940620422, + "learning_rate": 0.00016684865123437533, + "loss": 0.2551, + "step": 31880 + }, + { + "epoch": 0.49744181693392403, + "grad_norm": 2.9738593101501465, + "learning_rate": 0.00016683825211622055, + "loss": 0.4588, + "step": 31890 + }, + { + "epoch": 0.49759780370624573, + "grad_norm": 2.1061208248138428, + "learning_rate": 0.00016682785299806577, + "loss": 0.2302, + "step": 31900 + }, + { + "epoch": 0.49775379047856744, + "grad_norm": 1.5273017883300781, + "learning_rate": 0.000166817453879911, + "loss": 0.3038, + "step": 31910 + }, + { + "epoch": 0.49790977725088914, + "grad_norm": 8.00383472442627, + "learning_rate": 0.0001668070547617562, + "loss": 0.2682, + "step": 31920 + }, + { + "epoch": 0.49806576402321084, + "grad_norm": 2.956890344619751, + "learning_rate": 0.00016679665564360143, + "loss": 0.3565, + "step": 31930 + }, + { + "epoch": 0.49822175079553255, + "grad_norm": 1.0005966424942017, + "learning_rate": 0.00016678625652544665, + "loss": 0.203, + "step": 31940 + }, + { + "epoch": 0.49837773756785425, + "grad_norm": 1.8613911867141724, + "learning_rate": 0.00016677585740729187, + "loss": 0.2813, + "step": 31950 + }, + { + "epoch": 0.49853372434017595, + "grad_norm": 1.8933796882629395, + "learning_rate": 0.0001667654582891371, + "loss": 0.2361, + "step": 31960 + }, + { + "epoch": 0.49868971111249766, + "grad_norm": 0.12677444517612457, + "learning_rate": 0.0001667550591709823, + "loss": 0.1829, + "step": 31970 + }, + { + "epoch": 0.49884569788481936, + "grad_norm": 0.29421910643577576, + "learning_rate": 0.00016674466005282753, + "loss": 0.1087, + "step": 31980 + }, + { + "epoch": 0.49900168465714106, + "grad_norm": 1.9090936183929443, + "learning_rate": 0.00016673426093467275, + "loss": 0.2765, + "step": 31990 + }, + { + "epoch": 0.49915767142946277, + "grad_norm": 3.2520546913146973, + "learning_rate": 0.00016672386181651797, + "loss": 0.3858, + "step": 32000 + }, + { + "epoch": 0.49931365820178447, + "grad_norm": 0.9081943035125732, + "learning_rate": 0.00016671346269836319, + "loss": 0.1688, + "step": 32010 + }, + { + "epoch": 0.4994696449741062, + "grad_norm": 0.6389210224151611, + "learning_rate": 0.0001667030635802084, + "loss": 0.1244, + "step": 32020 + }, + { + "epoch": 0.4996256317464279, + "grad_norm": 0.7104432582855225, + "learning_rate": 0.00016669266446205363, + "loss": 0.2389, + "step": 32030 + }, + { + "epoch": 0.49978161851874964, + "grad_norm": 0.04320487007498741, + "learning_rate": 0.00016668226534389885, + "loss": 0.1421, + "step": 32040 + }, + { + "epoch": 0.49993760529107134, + "grad_norm": 1.359023094177246, + "learning_rate": 0.00016667186622574406, + "loss": 0.3741, + "step": 32050 + }, + { + "epoch": 0.500093592063393, + "grad_norm": 1.0862740278244019, + "learning_rate": 0.00016666146710758928, + "loss": 0.2905, + "step": 32060 + }, + { + "epoch": 0.5002495788357147, + "grad_norm": 3.362323045730591, + "learning_rate": 0.0001666510679894345, + "loss": 0.3792, + "step": 32070 + }, + { + "epoch": 0.5004055656080364, + "grad_norm": 0.2101036161184311, + "learning_rate": 0.00016664066887127972, + "loss": 0.1497, + "step": 32080 + }, + { + "epoch": 0.5005615523803582, + "grad_norm": 3.6670923233032227, + "learning_rate": 0.00016663026975312494, + "loss": 0.4363, + "step": 32090 + }, + { + "epoch": 0.5007175391526798, + "grad_norm": 1.1535643339157104, + "learning_rate": 0.00016661987063497016, + "loss": 0.4215, + "step": 32100 + }, + { + "epoch": 0.5008735259250016, + "grad_norm": 1.2060534954071045, + "learning_rate": 0.00016660947151681538, + "loss": 0.2109, + "step": 32110 + }, + { + "epoch": 0.5010295126973233, + "grad_norm": 1.1457877159118652, + "learning_rate": 0.0001665990723986606, + "loss": 0.2375, + "step": 32120 + }, + { + "epoch": 0.501185499469645, + "grad_norm": 1.1202179193496704, + "learning_rate": 0.00016658867328050582, + "loss": 0.1471, + "step": 32130 + }, + { + "epoch": 0.5013414862419667, + "grad_norm": 2.0000226497650146, + "learning_rate": 0.00016657827416235104, + "loss": 0.443, + "step": 32140 + }, + { + "epoch": 0.5014974730142884, + "grad_norm": 1.8489691019058228, + "learning_rate": 0.00016656787504419626, + "loss": 0.3755, + "step": 32150 + }, + { + "epoch": 0.5016534597866101, + "grad_norm": 1.055222511291504, + "learning_rate": 0.00016655747592604148, + "loss": 0.3125, + "step": 32160 + }, + { + "epoch": 0.5018094465589318, + "grad_norm": 3.8257789611816406, + "learning_rate": 0.0001665470768078867, + "loss": 0.1329, + "step": 32170 + }, + { + "epoch": 0.5019654333312535, + "grad_norm": 7.680811882019043, + "learning_rate": 0.00016653667768973192, + "loss": 0.3162, + "step": 32180 + }, + { + "epoch": 0.5021214201035752, + "grad_norm": 0.7791313529014587, + "learning_rate": 0.00016652627857157714, + "loss": 0.3993, + "step": 32190 + }, + { + "epoch": 0.502277406875897, + "grad_norm": 1.0188349485397339, + "learning_rate": 0.00016651587945342236, + "loss": 0.2772, + "step": 32200 + }, + { + "epoch": 0.5024333936482186, + "grad_norm": 1.3411445617675781, + "learning_rate": 0.00016650548033526758, + "loss": 0.3259, + "step": 32210 + }, + { + "epoch": 0.5025893804205404, + "grad_norm": 0.952900230884552, + "learning_rate": 0.0001664950812171128, + "loss": 0.3012, + "step": 32220 + }, + { + "epoch": 0.502745367192862, + "grad_norm": 3.059582233428955, + "learning_rate": 0.00016648468209895802, + "loss": 0.1581, + "step": 32230 + }, + { + "epoch": 0.5029013539651838, + "grad_norm": 3.8336679935455322, + "learning_rate": 0.00016647428298080324, + "loss": 0.2566, + "step": 32240 + }, + { + "epoch": 0.5030573407375054, + "grad_norm": 0.8763723969459534, + "learning_rate": 0.00016646388386264846, + "loss": 0.1746, + "step": 32250 + }, + { + "epoch": 0.5032133275098272, + "grad_norm": 5.046780109405518, + "learning_rate": 0.00016645348474449368, + "loss": 0.3014, + "step": 32260 + }, + { + "epoch": 0.5033693142821489, + "grad_norm": 0.046015236526727676, + "learning_rate": 0.0001664430856263389, + "loss": 0.1704, + "step": 32270 + }, + { + "epoch": 0.5035253010544706, + "grad_norm": 2.447784423828125, + "learning_rate": 0.00016643268650818412, + "loss": 0.4511, + "step": 32280 + }, + { + "epoch": 0.5036812878267923, + "grad_norm": 1.1856467723846436, + "learning_rate": 0.00016642228739002934, + "loss": 0.2444, + "step": 32290 + }, + { + "epoch": 0.503837274599114, + "grad_norm": 2.384140968322754, + "learning_rate": 0.00016641188827187456, + "loss": 0.5674, + "step": 32300 + }, + { + "epoch": 0.5039932613714357, + "grad_norm": 1.3639636039733887, + "learning_rate": 0.00016640148915371978, + "loss": 0.3968, + "step": 32310 + }, + { + "epoch": 0.5041492481437574, + "grad_norm": 1.4687482118606567, + "learning_rate": 0.000166391090035565, + "loss": 0.4899, + "step": 32320 + }, + { + "epoch": 0.5043052349160791, + "grad_norm": 1.5374367237091064, + "learning_rate": 0.00016638069091741021, + "loss": 0.3705, + "step": 32330 + }, + { + "epoch": 0.5044612216884008, + "grad_norm": 0.6463915109634399, + "learning_rate": 0.00016637029179925543, + "loss": 0.1967, + "step": 32340 + }, + { + "epoch": 0.5046172084607226, + "grad_norm": 1.7496861219406128, + "learning_rate": 0.00016635989268110065, + "loss": 0.4982, + "step": 32350 + }, + { + "epoch": 0.5047731952330442, + "grad_norm": 0.884661078453064, + "learning_rate": 0.00016634949356294587, + "loss": 0.2253, + "step": 32360 + }, + { + "epoch": 0.504929182005366, + "grad_norm": 1.5010557174682617, + "learning_rate": 0.0001663390944447911, + "loss": 0.3079, + "step": 32370 + }, + { + "epoch": 0.5050851687776876, + "grad_norm": 1.2658463716506958, + "learning_rate": 0.0001663286953266363, + "loss": 0.1173, + "step": 32380 + }, + { + "epoch": 0.5052411555500094, + "grad_norm": 0.18316659331321716, + "learning_rate": 0.00016631829620848153, + "loss": 0.2566, + "step": 32390 + }, + { + "epoch": 0.505397142322331, + "grad_norm": 2.102644681930542, + "learning_rate": 0.00016630789709032675, + "loss": 0.277, + "step": 32400 + }, + { + "epoch": 0.5055531290946528, + "grad_norm": 0.7332143187522888, + "learning_rate": 0.00016629749797217197, + "loss": 0.2754, + "step": 32410 + }, + { + "epoch": 0.5057091158669745, + "grad_norm": 1.9821066856384277, + "learning_rate": 0.0001662870988540172, + "loss": 0.3465, + "step": 32420 + }, + { + "epoch": 0.5058651026392962, + "grad_norm": 0.02897103875875473, + "learning_rate": 0.0001662766997358624, + "loss": 0.1438, + "step": 32430 + }, + { + "epoch": 0.5060210894116179, + "grad_norm": 0.3612467050552368, + "learning_rate": 0.00016626630061770763, + "loss": 0.3501, + "step": 32440 + }, + { + "epoch": 0.5061770761839396, + "grad_norm": 1.3284744024276733, + "learning_rate": 0.00016625590149955282, + "loss": 0.3822, + "step": 32450 + }, + { + "epoch": 0.5063330629562613, + "grad_norm": 2.2305402755737305, + "learning_rate": 0.00016624550238139807, + "loss": 0.2969, + "step": 32460 + }, + { + "epoch": 0.506489049728583, + "grad_norm": 1.6659493446350098, + "learning_rate": 0.00016623510326324326, + "loss": 0.2159, + "step": 32470 + }, + { + "epoch": 0.5066450365009048, + "grad_norm": 1.0163401365280151, + "learning_rate": 0.0001662247041450885, + "loss": 0.3209, + "step": 32480 + }, + { + "epoch": 0.5068010232732264, + "grad_norm": 0.9215751886367798, + "learning_rate": 0.0001662143050269337, + "loss": 0.2858, + "step": 32490 + }, + { + "epoch": 0.5069570100455482, + "grad_norm": 0.17924760282039642, + "learning_rate": 0.00016620390590877895, + "loss": 0.1559, + "step": 32500 + }, + { + "epoch": 0.5071129968178698, + "grad_norm": 1.0075199604034424, + "learning_rate": 0.00016619350679062414, + "loss": 0.2569, + "step": 32510 + }, + { + "epoch": 0.5072689835901916, + "grad_norm": 1.3489066362380981, + "learning_rate": 0.0001661831076724694, + "loss": 0.2565, + "step": 32520 + }, + { + "epoch": 0.5074249703625132, + "grad_norm": 0.2534058094024658, + "learning_rate": 0.00016617270855431458, + "loss": 0.2292, + "step": 32530 + }, + { + "epoch": 0.507580957134835, + "grad_norm": 0.31723281741142273, + "learning_rate": 0.00016616230943615983, + "loss": 0.1467, + "step": 32540 + }, + { + "epoch": 0.5077369439071566, + "grad_norm": 0.29659849405288696, + "learning_rate": 0.00016615191031800502, + "loss": 0.2768, + "step": 32550 + }, + { + "epoch": 0.5078929306794784, + "grad_norm": 1.5246435403823853, + "learning_rate": 0.00016614151119985027, + "loss": 0.216, + "step": 32560 + }, + { + "epoch": 0.5080489174518001, + "grad_norm": 0.6026854515075684, + "learning_rate": 0.00016613111208169546, + "loss": 0.2277, + "step": 32570 + }, + { + "epoch": 0.5082049042241218, + "grad_norm": 0.41994890570640564, + "learning_rate": 0.0001661207129635407, + "loss": 0.2433, + "step": 32580 + }, + { + "epoch": 0.5083608909964435, + "grad_norm": 2.084599733352661, + "learning_rate": 0.00016611031384538593, + "loss": 0.4081, + "step": 32590 + }, + { + "epoch": 0.5085168777687652, + "grad_norm": 0.2945229709148407, + "learning_rate": 0.00016609991472723115, + "loss": 0.1329, + "step": 32600 + }, + { + "epoch": 0.508672864541087, + "grad_norm": 1.4981423616409302, + "learning_rate": 0.00016608951560907636, + "loss": 0.2855, + "step": 32610 + }, + { + "epoch": 0.5088288513134086, + "grad_norm": 0.961737871170044, + "learning_rate": 0.00016607911649092158, + "loss": 0.1654, + "step": 32620 + }, + { + "epoch": 0.5089848380857304, + "grad_norm": 1.49734628200531, + "learning_rate": 0.0001660687173727668, + "loss": 0.2081, + "step": 32630 + }, + { + "epoch": 0.509140824858052, + "grad_norm": 2.538135051727295, + "learning_rate": 0.00016605831825461202, + "loss": 0.3023, + "step": 32640 + }, + { + "epoch": 0.5092968116303738, + "grad_norm": 0.2981272041797638, + "learning_rate": 0.00016604791913645724, + "loss": 0.214, + "step": 32650 + }, + { + "epoch": 0.5094527984026954, + "grad_norm": 1.8863650560379028, + "learning_rate": 0.00016603752001830246, + "loss": 0.3363, + "step": 32660 + }, + { + "epoch": 0.5096087851750172, + "grad_norm": 0.022720983251929283, + "learning_rate": 0.00016602712090014768, + "loss": 0.2463, + "step": 32670 + }, + { + "epoch": 0.5097647719473388, + "grad_norm": 1.154977798461914, + "learning_rate": 0.0001660167217819929, + "loss": 0.4484, + "step": 32680 + }, + { + "epoch": 0.5099207587196606, + "grad_norm": 0.9470499157905579, + "learning_rate": 0.00016600632266383812, + "loss": 0.1196, + "step": 32690 + }, + { + "epoch": 0.5100767454919822, + "grad_norm": 0.19955401122570038, + "learning_rate": 0.00016599592354568334, + "loss": 0.1762, + "step": 32700 + }, + { + "epoch": 0.510232732264304, + "grad_norm": 0.4098282754421234, + "learning_rate": 0.00016598552442752856, + "loss": 0.2739, + "step": 32710 + }, + { + "epoch": 0.5103887190366257, + "grad_norm": 0.21697258949279785, + "learning_rate": 0.00016597512530937378, + "loss": 0.2555, + "step": 32720 + }, + { + "epoch": 0.5105447058089474, + "grad_norm": 4.894479751586914, + "learning_rate": 0.000165964726191219, + "loss": 0.3137, + "step": 32730 + }, + { + "epoch": 0.5107006925812692, + "grad_norm": 0.8396711349487305, + "learning_rate": 0.00016595432707306422, + "loss": 0.3105, + "step": 32740 + }, + { + "epoch": 0.5108566793535908, + "grad_norm": 1.5494674444198608, + "learning_rate": 0.00016594392795490944, + "loss": 0.2271, + "step": 32750 + }, + { + "epoch": 0.5110126661259126, + "grad_norm": 1.3409888744354248, + "learning_rate": 0.00016593352883675466, + "loss": 0.1796, + "step": 32760 + }, + { + "epoch": 0.5111686528982342, + "grad_norm": 1.7590911388397217, + "learning_rate": 0.00016592312971859988, + "loss": 0.2185, + "step": 32770 + }, + { + "epoch": 0.511324639670556, + "grad_norm": 0.7350002527236938, + "learning_rate": 0.0001659127306004451, + "loss": 0.1911, + "step": 32780 + }, + { + "epoch": 0.5114806264428776, + "grad_norm": 0.7176411747932434, + "learning_rate": 0.00016590233148229032, + "loss": 0.6918, + "step": 32790 + }, + { + "epoch": 0.5116366132151994, + "grad_norm": 2.5005321502685547, + "learning_rate": 0.00016589193236413554, + "loss": 0.2329, + "step": 32800 + }, + { + "epoch": 0.511792599987521, + "grad_norm": 1.3916250467300415, + "learning_rate": 0.00016588153324598076, + "loss": 0.2102, + "step": 32810 + }, + { + "epoch": 0.5119485867598428, + "grad_norm": 0.721671998500824, + "learning_rate": 0.00016587113412782598, + "loss": 0.2965, + "step": 32820 + }, + { + "epoch": 0.5121045735321644, + "grad_norm": 0.6120114326477051, + "learning_rate": 0.0001658607350096712, + "loss": 0.2661, + "step": 32830 + }, + { + "epoch": 0.5122605603044862, + "grad_norm": 0.29123347997665405, + "learning_rate": 0.00016585033589151642, + "loss": 0.2455, + "step": 32840 + }, + { + "epoch": 0.5124165470768078, + "grad_norm": 2.7300868034362793, + "learning_rate": 0.00016583993677336164, + "loss": 0.2049, + "step": 32850 + }, + { + "epoch": 0.5125725338491296, + "grad_norm": 1.801928997039795, + "learning_rate": 0.00016582953765520686, + "loss": 0.1725, + "step": 32860 + }, + { + "epoch": 0.5127285206214514, + "grad_norm": 0.6406471729278564, + "learning_rate": 0.00016581913853705208, + "loss": 0.164, + "step": 32870 + }, + { + "epoch": 0.512884507393773, + "grad_norm": 0.9627325534820557, + "learning_rate": 0.0001658087394188973, + "loss": 0.1728, + "step": 32880 + }, + { + "epoch": 0.5130404941660948, + "grad_norm": 1.5446962118148804, + "learning_rate": 0.00016579834030074251, + "loss": 0.2345, + "step": 32890 + }, + { + "epoch": 0.5131964809384164, + "grad_norm": 0.28092706203460693, + "learning_rate": 0.0001657879411825877, + "loss": 0.2124, + "step": 32900 + }, + { + "epoch": 0.5133524677107382, + "grad_norm": 1.7751421928405762, + "learning_rate": 0.00016577754206443295, + "loss": 0.2367, + "step": 32910 + }, + { + "epoch": 0.5135084544830598, + "grad_norm": 1.6964634656906128, + "learning_rate": 0.00016576714294627815, + "loss": 0.2361, + "step": 32920 + }, + { + "epoch": 0.5136644412553816, + "grad_norm": 0.06759823113679886, + "learning_rate": 0.0001657567438281234, + "loss": 0.2329, + "step": 32930 + }, + { + "epoch": 0.5138204280277032, + "grad_norm": 2.2519280910491943, + "learning_rate": 0.00016574634470996859, + "loss": 0.1079, + "step": 32940 + }, + { + "epoch": 0.513976414800025, + "grad_norm": 3.167576551437378, + "learning_rate": 0.00016573594559181383, + "loss": 0.1764, + "step": 32950 + }, + { + "epoch": 0.5141324015723466, + "grad_norm": 1.6734397411346436, + "learning_rate": 0.00016572554647365903, + "loss": 0.2878, + "step": 32960 + }, + { + "epoch": 0.5142883883446684, + "grad_norm": 0.35627374053001404, + "learning_rate": 0.00016571514735550427, + "loss": 0.2337, + "step": 32970 + }, + { + "epoch": 0.51444437511699, + "grad_norm": 1.6321039199829102, + "learning_rate": 0.00016570474823734946, + "loss": 0.2551, + "step": 32980 + }, + { + "epoch": 0.5146003618893118, + "grad_norm": 1.352982997894287, + "learning_rate": 0.0001656943491191947, + "loss": 0.4467, + "step": 32990 + }, + { + "epoch": 0.5147563486616334, + "grad_norm": 0.3558129072189331, + "learning_rate": 0.0001656839500010399, + "loss": 0.1913, + "step": 33000 + }, + { + "epoch": 0.5149123354339552, + "grad_norm": 0.041972286999225616, + "learning_rate": 0.00016567355088288515, + "loss": 0.1704, + "step": 33010 + }, + { + "epoch": 0.515068322206277, + "grad_norm": 1.4168939590454102, + "learning_rate": 0.00016566315176473034, + "loss": 0.4866, + "step": 33020 + }, + { + "epoch": 0.5152243089785986, + "grad_norm": 3.0042929649353027, + "learning_rate": 0.0001656527526465756, + "loss": 0.2473, + "step": 33030 + }, + { + "epoch": 0.5153802957509204, + "grad_norm": 1.027274489402771, + "learning_rate": 0.00016564235352842078, + "loss": 0.436, + "step": 33040 + }, + { + "epoch": 0.515536282523242, + "grad_norm": 1.2330195903778076, + "learning_rate": 0.00016563195441026603, + "loss": 0.4615, + "step": 33050 + }, + { + "epoch": 0.5156922692955638, + "grad_norm": 2.7145066261291504, + "learning_rate": 0.00016562155529211122, + "loss": 0.278, + "step": 33060 + }, + { + "epoch": 0.5158482560678854, + "grad_norm": 1.1273893117904663, + "learning_rate": 0.00016561115617395647, + "loss": 0.3767, + "step": 33070 + }, + { + "epoch": 0.5160042428402072, + "grad_norm": 2.466265916824341, + "learning_rate": 0.00016560075705580166, + "loss": 0.1879, + "step": 33080 + }, + { + "epoch": 0.5161602296125288, + "grad_norm": 0.3759515583515167, + "learning_rate": 0.0001655903579376469, + "loss": 0.4004, + "step": 33090 + }, + { + "epoch": 0.5163162163848506, + "grad_norm": 1.8821336030960083, + "learning_rate": 0.0001655799588194921, + "loss": 0.1622, + "step": 33100 + }, + { + "epoch": 0.5164722031571722, + "grad_norm": 0.8162096738815308, + "learning_rate": 0.00016556955970133735, + "loss": 0.1341, + "step": 33110 + }, + { + "epoch": 0.516628189929494, + "grad_norm": 0.008498801849782467, + "learning_rate": 0.00016555916058318254, + "loss": 0.2546, + "step": 33120 + }, + { + "epoch": 0.5167841767018156, + "grad_norm": 0.17536675930023193, + "learning_rate": 0.00016554876146502779, + "loss": 0.2934, + "step": 33130 + }, + { + "epoch": 0.5169401634741374, + "grad_norm": 3.539898633956909, + "learning_rate": 0.00016553836234687298, + "loss": 0.4959, + "step": 33140 + }, + { + "epoch": 0.517096150246459, + "grad_norm": 2.190887212753296, + "learning_rate": 0.00016552796322871823, + "loss": 0.2914, + "step": 33150 + }, + { + "epoch": 0.5172521370187808, + "grad_norm": 1.6388943195343018, + "learning_rate": 0.00016551756411056342, + "loss": 0.1317, + "step": 33160 + }, + { + "epoch": 0.5174081237911026, + "grad_norm": 0.33063387870788574, + "learning_rate": 0.00016550716499240866, + "loss": 0.4258, + "step": 33170 + }, + { + "epoch": 0.5175641105634242, + "grad_norm": 0.5168288946151733, + "learning_rate": 0.00016549676587425386, + "loss": 0.2548, + "step": 33180 + }, + { + "epoch": 0.517720097335746, + "grad_norm": 3.9570281505584717, + "learning_rate": 0.0001654863667560991, + "loss": 0.3179, + "step": 33190 + }, + { + "epoch": 0.5178760841080676, + "grad_norm": 1.127255916595459, + "learning_rate": 0.0001654759676379443, + "loss": 0.233, + "step": 33200 + }, + { + "epoch": 0.5180320708803894, + "grad_norm": 0.36299410462379456, + "learning_rate": 0.00016546556851978954, + "loss": 0.4614, + "step": 33210 + }, + { + "epoch": 0.518188057652711, + "grad_norm": 2.914668321609497, + "learning_rate": 0.00016545516940163474, + "loss": 0.1603, + "step": 33220 + }, + { + "epoch": 0.5183440444250328, + "grad_norm": 6.991094589233398, + "learning_rate": 0.00016544477028347998, + "loss": 0.2259, + "step": 33230 + }, + { + "epoch": 0.5185000311973544, + "grad_norm": 2.0594241619110107, + "learning_rate": 0.00016543437116532517, + "loss": 0.1885, + "step": 33240 + }, + { + "epoch": 0.5186560179696762, + "grad_norm": 0.5594693422317505, + "learning_rate": 0.00016542397204717042, + "loss": 0.4333, + "step": 33250 + }, + { + "epoch": 0.5188120047419978, + "grad_norm": 4.314854621887207, + "learning_rate": 0.00016541357292901561, + "loss": 0.2103, + "step": 33260 + }, + { + "epoch": 0.5189679915143196, + "grad_norm": 0.5609000325202942, + "learning_rate": 0.00016540317381086086, + "loss": 0.2498, + "step": 33270 + }, + { + "epoch": 0.5191239782866413, + "grad_norm": 3.680088758468628, + "learning_rate": 0.00016539277469270605, + "loss": 0.3465, + "step": 33280 + }, + { + "epoch": 0.519279965058963, + "grad_norm": 1.4993672370910645, + "learning_rate": 0.0001653823755745513, + "loss": 0.3272, + "step": 33290 + }, + { + "epoch": 0.5194359518312847, + "grad_norm": 0.18005084991455078, + "learning_rate": 0.0001653719764563965, + "loss": 0.1706, + "step": 33300 + }, + { + "epoch": 0.5195919386036064, + "grad_norm": 0.5105524659156799, + "learning_rate": 0.00016536157733824174, + "loss": 0.0937, + "step": 33310 + }, + { + "epoch": 0.5197479253759281, + "grad_norm": 1.2806589603424072, + "learning_rate": 0.00016535117822008693, + "loss": 0.1578, + "step": 33320 + }, + { + "epoch": 0.5199039121482498, + "grad_norm": 0.22015990316867828, + "learning_rate": 0.00016534077910193218, + "loss": 0.2081, + "step": 33330 + }, + { + "epoch": 0.5200598989205716, + "grad_norm": 0.34638991951942444, + "learning_rate": 0.00016533037998377737, + "loss": 0.2389, + "step": 33340 + }, + { + "epoch": 0.5202158856928932, + "grad_norm": 0.8709302544593811, + "learning_rate": 0.0001653199808656226, + "loss": 0.2112, + "step": 33350 + }, + { + "epoch": 0.520371872465215, + "grad_norm": 1.2309521436691284, + "learning_rate": 0.0001653095817474678, + "loss": 0.3448, + "step": 33360 + }, + { + "epoch": 0.5205278592375366, + "grad_norm": 1.3155115842819214, + "learning_rate": 0.00016529918262931303, + "loss": 0.1572, + "step": 33370 + }, + { + "epoch": 0.5206838460098584, + "grad_norm": 0.358751118183136, + "learning_rate": 0.00016528878351115825, + "loss": 0.3011, + "step": 33380 + }, + { + "epoch": 0.52083983278218, + "grad_norm": 3.800628662109375, + "learning_rate": 0.00016527838439300347, + "loss": 0.458, + "step": 33390 + }, + { + "epoch": 0.5209958195545018, + "grad_norm": 1.1253124475479126, + "learning_rate": 0.0001652679852748487, + "loss": 0.1238, + "step": 33400 + }, + { + "epoch": 0.5211518063268235, + "grad_norm": 0.46979108452796936, + "learning_rate": 0.0001652575861566939, + "loss": 0.0745, + "step": 33410 + }, + { + "epoch": 0.5213077930991452, + "grad_norm": 7.921492576599121, + "learning_rate": 0.00016524718703853913, + "loss": 0.4004, + "step": 33420 + }, + { + "epoch": 0.5214637798714669, + "grad_norm": 1.643127679824829, + "learning_rate": 0.00016523678792038435, + "loss": 0.2858, + "step": 33430 + }, + { + "epoch": 0.5216197666437886, + "grad_norm": 7.096631050109863, + "learning_rate": 0.0001652263888022296, + "loss": 0.4735, + "step": 33440 + }, + { + "epoch": 0.5217757534161103, + "grad_norm": 0.7280219197273254, + "learning_rate": 0.0001652159896840748, + "loss": 0.1841, + "step": 33450 + }, + { + "epoch": 0.521931740188432, + "grad_norm": 0.9415522813796997, + "learning_rate": 0.00016520559056592003, + "loss": 0.2169, + "step": 33460 + }, + { + "epoch": 0.5220877269607537, + "grad_norm": 2.264862060546875, + "learning_rate": 0.00016519519144776523, + "loss": 0.1902, + "step": 33470 + }, + { + "epoch": 0.5222437137330754, + "grad_norm": 2.4566822052001953, + "learning_rate": 0.00016518479232961047, + "loss": 0.3329, + "step": 33480 + }, + { + "epoch": 0.5223997005053972, + "grad_norm": 0.7330378890037537, + "learning_rate": 0.00016517439321145567, + "loss": 0.2341, + "step": 33490 + }, + { + "epoch": 0.5225556872777188, + "grad_norm": 1.6409542560577393, + "learning_rate": 0.0001651639940933009, + "loss": 0.3077, + "step": 33500 + }, + { + "epoch": 0.5227116740500406, + "grad_norm": 1.0647432804107666, + "learning_rate": 0.0001651535949751461, + "loss": 0.1935, + "step": 33510 + }, + { + "epoch": 0.5228676608223622, + "grad_norm": 0.507420539855957, + "learning_rate": 0.00016514319585699135, + "loss": 0.1314, + "step": 33520 + }, + { + "epoch": 0.523023647594684, + "grad_norm": 1.0064164400100708, + "learning_rate": 0.00016513279673883654, + "loss": 0.344, + "step": 33530 + }, + { + "epoch": 0.5231796343670057, + "grad_norm": 1.1936209201812744, + "learning_rate": 0.0001651223976206818, + "loss": 0.1948, + "step": 33540 + }, + { + "epoch": 0.5233356211393274, + "grad_norm": 0.04992926865816116, + "learning_rate": 0.00016511199850252698, + "loss": 0.1363, + "step": 33550 + }, + { + "epoch": 0.5234916079116491, + "grad_norm": 1.8256275653839111, + "learning_rate": 0.00016510159938437223, + "loss": 0.443, + "step": 33560 + }, + { + "epoch": 0.5236475946839708, + "grad_norm": 4.914908409118652, + "learning_rate": 0.00016509120026621742, + "loss": 0.4654, + "step": 33570 + }, + { + "epoch": 0.5238035814562925, + "grad_norm": 1.1625486612319946, + "learning_rate": 0.00016508080114806267, + "loss": 0.1974, + "step": 33580 + }, + { + "epoch": 0.5239595682286142, + "grad_norm": 6.6192522048950195, + "learning_rate": 0.00016507040202990786, + "loss": 0.2131, + "step": 33590 + }, + { + "epoch": 0.5241155550009359, + "grad_norm": 0.7369065880775452, + "learning_rate": 0.0001650600029117531, + "loss": 0.2371, + "step": 33600 + }, + { + "epoch": 0.5242715417732576, + "grad_norm": 1.5238152742385864, + "learning_rate": 0.0001650496037935983, + "loss": 0.1557, + "step": 33610 + }, + { + "epoch": 0.5244275285455793, + "grad_norm": 1.0418007373809814, + "learning_rate": 0.00016503920467544355, + "loss": 0.1878, + "step": 33620 + }, + { + "epoch": 0.524583515317901, + "grad_norm": 0.790117084980011, + "learning_rate": 0.00016502880555728874, + "loss": 0.2195, + "step": 33630 + }, + { + "epoch": 0.5247395020902228, + "grad_norm": 1.6712257862091064, + "learning_rate": 0.000165018406439134, + "loss": 0.1602, + "step": 33640 + }, + { + "epoch": 0.5248954888625444, + "grad_norm": 0.19236230850219727, + "learning_rate": 0.00016500800732097918, + "loss": 0.2526, + "step": 33650 + }, + { + "epoch": 0.5250514756348662, + "grad_norm": 1.3519701957702637, + "learning_rate": 0.00016499760820282443, + "loss": 0.4686, + "step": 33660 + }, + { + "epoch": 0.5252074624071879, + "grad_norm": 1.694342851638794, + "learning_rate": 0.00016498720908466962, + "loss": 0.1859, + "step": 33670 + }, + { + "epoch": 0.5253634491795096, + "grad_norm": 5.225239276885986, + "learning_rate": 0.00016497680996651487, + "loss": 0.2944, + "step": 33680 + }, + { + "epoch": 0.5255194359518313, + "grad_norm": 2.0208842754364014, + "learning_rate": 0.00016496641084836006, + "loss": 0.2421, + "step": 33690 + }, + { + "epoch": 0.525675422724153, + "grad_norm": 0.7954996228218079, + "learning_rate": 0.0001649560117302053, + "loss": 0.2853, + "step": 33700 + }, + { + "epoch": 0.5258314094964747, + "grad_norm": 2.296086072921753, + "learning_rate": 0.0001649456126120505, + "loss": 0.1333, + "step": 33710 + }, + { + "epoch": 0.5259873962687964, + "grad_norm": 1.1779128313064575, + "learning_rate": 0.00016493521349389574, + "loss": 0.1066, + "step": 33720 + }, + { + "epoch": 0.5261433830411181, + "grad_norm": 0.1756065934896469, + "learning_rate": 0.00016492481437574094, + "loss": 0.1352, + "step": 33730 + }, + { + "epoch": 0.5262993698134398, + "grad_norm": 0.13100725412368774, + "learning_rate": 0.00016491441525758618, + "loss": 0.2399, + "step": 33740 + }, + { + "epoch": 0.5264553565857615, + "grad_norm": 5.532008171081543, + "learning_rate": 0.00016490401613943138, + "loss": 0.2896, + "step": 33750 + }, + { + "epoch": 0.5266113433580832, + "grad_norm": 1.319886565208435, + "learning_rate": 0.00016489361702127662, + "loss": 0.3275, + "step": 33760 + }, + { + "epoch": 0.5267673301304049, + "grad_norm": 1.5550974607467651, + "learning_rate": 0.00016488321790312182, + "loss": 0.2677, + "step": 33770 + }, + { + "epoch": 0.5269233169027266, + "grad_norm": 1.8936737775802612, + "learning_rate": 0.00016487281878496706, + "loss": 0.1955, + "step": 33780 + }, + { + "epoch": 0.5270793036750484, + "grad_norm": 0.3653401732444763, + "learning_rate": 0.00016486241966681226, + "loss": 0.0723, + "step": 33790 + }, + { + "epoch": 0.52723529044737, + "grad_norm": 2.861341714859009, + "learning_rate": 0.00016485202054865747, + "loss": 0.2412, + "step": 33800 + }, + { + "epoch": 0.5273912772196918, + "grad_norm": 1.5291428565979004, + "learning_rate": 0.0001648416214305027, + "loss": 0.0871, + "step": 33810 + }, + { + "epoch": 0.5275472639920135, + "grad_norm": 1.0372581481933594, + "learning_rate": 0.00016483122231234791, + "loss": 0.4705, + "step": 33820 + }, + { + "epoch": 0.5277032507643352, + "grad_norm": 1.1943141222000122, + "learning_rate": 0.00016482082319419313, + "loss": 0.2848, + "step": 33830 + }, + { + "epoch": 0.5278592375366569, + "grad_norm": 1.9008225202560425, + "learning_rate": 0.00016481042407603835, + "loss": 0.139, + "step": 33840 + }, + { + "epoch": 0.5280152243089786, + "grad_norm": 2.132089138031006, + "learning_rate": 0.00016480002495788357, + "loss": 0.2119, + "step": 33850 + }, + { + "epoch": 0.5281712110813003, + "grad_norm": 0.24524426460266113, + "learning_rate": 0.0001647896258397288, + "loss": 0.1643, + "step": 33860 + }, + { + "epoch": 0.528327197853622, + "grad_norm": 1.6469637155532837, + "learning_rate": 0.000164779226721574, + "loss": 0.363, + "step": 33870 + }, + { + "epoch": 0.5284831846259437, + "grad_norm": 0.8767328858375549, + "learning_rate": 0.00016476882760341923, + "loss": 0.2632, + "step": 33880 + }, + { + "epoch": 0.5286391713982654, + "grad_norm": 0.06347586214542389, + "learning_rate": 0.00016475842848526445, + "loss": 0.3204, + "step": 33890 + }, + { + "epoch": 0.5287951581705871, + "grad_norm": 0.09782540798187256, + "learning_rate": 0.00016474802936710967, + "loss": 0.2211, + "step": 33900 + }, + { + "epoch": 0.5289511449429088, + "grad_norm": 3.2998859882354736, + "learning_rate": 0.0001647376302489549, + "loss": 0.265, + "step": 33910 + }, + { + "epoch": 0.5291071317152305, + "grad_norm": 0.43594226241111755, + "learning_rate": 0.0001647272311308001, + "loss": 0.194, + "step": 33920 + }, + { + "epoch": 0.5292631184875523, + "grad_norm": 1.5166605710983276, + "learning_rate": 0.00016471683201264533, + "loss": 0.2675, + "step": 33930 + }, + { + "epoch": 0.529419105259874, + "grad_norm": 0.6056640148162842, + "learning_rate": 0.00016470643289449055, + "loss": 0.0576, + "step": 33940 + }, + { + "epoch": 0.5295750920321957, + "grad_norm": 0.25410348176956177, + "learning_rate": 0.00016469603377633577, + "loss": 0.1124, + "step": 33950 + }, + { + "epoch": 0.5297310788045174, + "grad_norm": 1.770642876625061, + "learning_rate": 0.000164685634658181, + "loss": 0.3295, + "step": 33960 + }, + { + "epoch": 0.5298870655768391, + "grad_norm": 0.0607205331325531, + "learning_rate": 0.0001646752355400262, + "loss": 0.2369, + "step": 33970 + }, + { + "epoch": 0.5300430523491608, + "grad_norm": 0.5557095408439636, + "learning_rate": 0.00016466483642187143, + "loss": 0.1681, + "step": 33980 + }, + { + "epoch": 0.5301990391214825, + "grad_norm": 0.5192957520484924, + "learning_rate": 0.00016465443730371665, + "loss": 0.2649, + "step": 33990 + }, + { + "epoch": 0.5303550258938042, + "grad_norm": 0.04804835096001625, + "learning_rate": 0.00016464403818556187, + "loss": 0.2639, + "step": 34000 + }, + { + "epoch": 0.5305110126661259, + "grad_norm": 0.02673129364848137, + "learning_rate": 0.0001646336390674071, + "loss": 0.169, + "step": 34010 + }, + { + "epoch": 0.5306669994384476, + "grad_norm": 1.0084244012832642, + "learning_rate": 0.0001646232399492523, + "loss": 0.353, + "step": 34020 + }, + { + "epoch": 0.5308229862107693, + "grad_norm": 2.2202091217041016, + "learning_rate": 0.00016461284083109753, + "loss": 0.3206, + "step": 34030 + }, + { + "epoch": 0.530978972983091, + "grad_norm": 0.5573744773864746, + "learning_rate": 0.00016460244171294275, + "loss": 0.231, + "step": 34040 + }, + { + "epoch": 0.5311349597554127, + "grad_norm": 0.39700084924697876, + "learning_rate": 0.00016459204259478797, + "loss": 0.1412, + "step": 34050 + }, + { + "epoch": 0.5312909465277345, + "grad_norm": 2.582963228225708, + "learning_rate": 0.00016458164347663319, + "loss": 0.3334, + "step": 34060 + }, + { + "epoch": 0.5314469333000561, + "grad_norm": 0.22781169414520264, + "learning_rate": 0.0001645712443584784, + "loss": 0.2767, + "step": 34070 + }, + { + "epoch": 0.5316029200723779, + "grad_norm": 1.2051042318344116, + "learning_rate": 0.00016456084524032362, + "loss": 0.2045, + "step": 34080 + }, + { + "epoch": 0.5317589068446996, + "grad_norm": 0.42760100960731506, + "learning_rate": 0.00016455044612216884, + "loss": 0.1151, + "step": 34090 + }, + { + "epoch": 0.5319148936170213, + "grad_norm": 0.14440476894378662, + "learning_rate": 0.00016454004700401406, + "loss": 0.2386, + "step": 34100 + }, + { + "epoch": 0.532070880389343, + "grad_norm": 2.2777981758117676, + "learning_rate": 0.00016452964788585928, + "loss": 0.2009, + "step": 34110 + }, + { + "epoch": 0.5322268671616647, + "grad_norm": 0.9206979274749756, + "learning_rate": 0.0001645192487677045, + "loss": 0.2745, + "step": 34120 + }, + { + "epoch": 0.5323828539339864, + "grad_norm": 1.6947574615478516, + "learning_rate": 0.00016450884964954972, + "loss": 0.2584, + "step": 34130 + }, + { + "epoch": 0.5325388407063081, + "grad_norm": 0.401444673538208, + "learning_rate": 0.00016449845053139494, + "loss": 0.2218, + "step": 34140 + }, + { + "epoch": 0.5326948274786298, + "grad_norm": 0.08261553198099136, + "learning_rate": 0.00016448805141324016, + "loss": 0.2775, + "step": 34150 + }, + { + "epoch": 0.5328508142509515, + "grad_norm": 0.1017974391579628, + "learning_rate": 0.00016447765229508538, + "loss": 0.2095, + "step": 34160 + }, + { + "epoch": 0.5330068010232732, + "grad_norm": 1.3759571313858032, + "learning_rate": 0.0001644672531769306, + "loss": 0.2643, + "step": 34170 + }, + { + "epoch": 0.5331627877955949, + "grad_norm": 1.2654389142990112, + "learning_rate": 0.00016445685405877582, + "loss": 0.2949, + "step": 34180 + }, + { + "epoch": 0.5333187745679167, + "grad_norm": 1.5481843948364258, + "learning_rate": 0.00016444645494062104, + "loss": 0.1264, + "step": 34190 + }, + { + "epoch": 0.5334747613402383, + "grad_norm": 1.8094528913497925, + "learning_rate": 0.00016443605582246626, + "loss": 0.2727, + "step": 34200 + }, + { + "epoch": 0.5336307481125601, + "grad_norm": 2.224538564682007, + "learning_rate": 0.00016442565670431148, + "loss": 0.3096, + "step": 34210 + }, + { + "epoch": 0.5337867348848817, + "grad_norm": 0.6375226974487305, + "learning_rate": 0.0001644152575861567, + "loss": 0.2251, + "step": 34220 + }, + { + "epoch": 0.5339427216572035, + "grad_norm": 3.727106809616089, + "learning_rate": 0.00016440485846800192, + "loss": 0.4374, + "step": 34230 + }, + { + "epoch": 0.5340987084295252, + "grad_norm": 0.13345426321029663, + "learning_rate": 0.00016439445934984714, + "loss": 0.2011, + "step": 34240 + }, + { + "epoch": 0.5342546952018469, + "grad_norm": 2.1658668518066406, + "learning_rate": 0.00016438406023169236, + "loss": 0.2457, + "step": 34250 + }, + { + "epoch": 0.5344106819741686, + "grad_norm": 19.238407135009766, + "learning_rate": 0.00016437366111353758, + "loss": 0.2756, + "step": 34260 + }, + { + "epoch": 0.5345666687464903, + "grad_norm": 1.0292778015136719, + "learning_rate": 0.0001643632619953828, + "loss": 0.1646, + "step": 34270 + }, + { + "epoch": 0.534722655518812, + "grad_norm": 0.9372987747192383, + "learning_rate": 0.00016435286287722802, + "loss": 0.2762, + "step": 34280 + }, + { + "epoch": 0.5348786422911337, + "grad_norm": 0.3918002247810364, + "learning_rate": 0.00016434246375907324, + "loss": 0.1406, + "step": 34290 + }, + { + "epoch": 0.5350346290634554, + "grad_norm": 1.3518732786178589, + "learning_rate": 0.00016433206464091846, + "loss": 0.2992, + "step": 34300 + }, + { + "epoch": 0.5351906158357771, + "grad_norm": 0.73117595911026, + "learning_rate": 0.00016432166552276368, + "loss": 0.1172, + "step": 34310 + }, + { + "epoch": 0.5353466026080989, + "grad_norm": 0.06655958294868469, + "learning_rate": 0.0001643112664046089, + "loss": 0.1332, + "step": 34320 + }, + { + "epoch": 0.5355025893804205, + "grad_norm": 3.2587168216705322, + "learning_rate": 0.00016430086728645412, + "loss": 0.2052, + "step": 34330 + }, + { + "epoch": 0.5356585761527423, + "grad_norm": 2.6668431758880615, + "learning_rate": 0.00016429046816829934, + "loss": 0.1295, + "step": 34340 + }, + { + "epoch": 0.5358145629250639, + "grad_norm": 1.316080927848816, + "learning_rate": 0.00016428006905014456, + "loss": 0.2072, + "step": 34350 + }, + { + "epoch": 0.5359705496973857, + "grad_norm": 5.144528388977051, + "learning_rate": 0.00016426966993198977, + "loss": 0.1934, + "step": 34360 + }, + { + "epoch": 0.5361265364697073, + "grad_norm": 1.3529599905014038, + "learning_rate": 0.000164259270813835, + "loss": 0.2077, + "step": 34370 + }, + { + "epoch": 0.5362825232420291, + "grad_norm": 0.9132925868034363, + "learning_rate": 0.00016424887169568021, + "loss": 0.2404, + "step": 34380 + }, + { + "epoch": 0.5364385100143508, + "grad_norm": 1.520033597946167, + "learning_rate": 0.00016423847257752543, + "loss": 0.1765, + "step": 34390 + }, + { + "epoch": 0.5365944967866725, + "grad_norm": 2.343975305557251, + "learning_rate": 0.00016422807345937065, + "loss": 0.1959, + "step": 34400 + }, + { + "epoch": 0.5367504835589942, + "grad_norm": 0.5855118632316589, + "learning_rate": 0.00016421767434121587, + "loss": 0.0848, + "step": 34410 + }, + { + "epoch": 0.5369064703313159, + "grad_norm": 1.9108648300170898, + "learning_rate": 0.0001642072752230611, + "loss": 0.2406, + "step": 34420 + }, + { + "epoch": 0.5370624571036376, + "grad_norm": 1.0323792695999146, + "learning_rate": 0.0001641968761049063, + "loss": 0.2298, + "step": 34430 + }, + { + "epoch": 0.5372184438759593, + "grad_norm": 0.17435620725154877, + "learning_rate": 0.00016418647698675153, + "loss": 0.1396, + "step": 34440 + }, + { + "epoch": 0.537374430648281, + "grad_norm": 0.3526577353477478, + "learning_rate": 0.00016417607786859675, + "loss": 0.172, + "step": 34450 + }, + { + "epoch": 0.5375304174206027, + "grad_norm": 2.73685884475708, + "learning_rate": 0.00016416567875044197, + "loss": 0.1167, + "step": 34460 + }, + { + "epoch": 0.5376864041929245, + "grad_norm": 2.765693426132202, + "learning_rate": 0.0001641552796322872, + "loss": 0.1356, + "step": 34470 + }, + { + "epoch": 0.5378423909652461, + "grad_norm": 0.47628021240234375, + "learning_rate": 0.0001641448805141324, + "loss": 0.3104, + "step": 34480 + }, + { + "epoch": 0.5379983777375679, + "grad_norm": 1.1307590007781982, + "learning_rate": 0.00016413448139597763, + "loss": 0.3247, + "step": 34490 + }, + { + "epoch": 0.5381543645098895, + "grad_norm": 0.07136381417512894, + "learning_rate": 0.00016412408227782285, + "loss": 0.319, + "step": 34500 + }, + { + "epoch": 0.5383103512822113, + "grad_norm": 1.7850221395492554, + "learning_rate": 0.00016411368315966807, + "loss": 0.2515, + "step": 34510 + }, + { + "epoch": 0.5384663380545329, + "grad_norm": 1.3870742321014404, + "learning_rate": 0.0001641032840415133, + "loss": 0.256, + "step": 34520 + }, + { + "epoch": 0.5386223248268547, + "grad_norm": 1.1329221725463867, + "learning_rate": 0.0001640928849233585, + "loss": 0.2255, + "step": 34530 + }, + { + "epoch": 0.5387783115991764, + "grad_norm": 0.23236137628555298, + "learning_rate": 0.00016408248580520373, + "loss": 0.4053, + "step": 34540 + }, + { + "epoch": 0.5389342983714981, + "grad_norm": 3.350924253463745, + "learning_rate": 0.00016407208668704895, + "loss": 0.2816, + "step": 34550 + }, + { + "epoch": 0.5390902851438198, + "grad_norm": 1.7459170818328857, + "learning_rate": 0.00016406168756889417, + "loss": 0.0834, + "step": 34560 + }, + { + "epoch": 0.5392462719161415, + "grad_norm": 0.89351487159729, + "learning_rate": 0.0001640512884507394, + "loss": 0.1146, + "step": 34570 + }, + { + "epoch": 0.5394022586884633, + "grad_norm": 0.49410831928253174, + "learning_rate": 0.0001640408893325846, + "loss": 0.2319, + "step": 34580 + }, + { + "epoch": 0.5395582454607849, + "grad_norm": 0.009971237741410732, + "learning_rate": 0.00016403049021442983, + "loss": 0.1674, + "step": 34590 + }, + { + "epoch": 0.5397142322331067, + "grad_norm": 1.420511245727539, + "learning_rate": 0.00016402009109627505, + "loss": 0.1469, + "step": 34600 + }, + { + "epoch": 0.5398702190054283, + "grad_norm": 2.5686769485473633, + "learning_rate": 0.00016400969197812027, + "loss": 0.2607, + "step": 34610 + }, + { + "epoch": 0.5400262057777501, + "grad_norm": 2.5808913707733154, + "learning_rate": 0.00016399929285996549, + "loss": 0.4412, + "step": 34620 + }, + { + "epoch": 0.5401821925500717, + "grad_norm": 0.2778591811656952, + "learning_rate": 0.0001639888937418107, + "loss": 0.1563, + "step": 34630 + }, + { + "epoch": 0.5403381793223935, + "grad_norm": 1.6448099613189697, + "learning_rate": 0.00016397849462365592, + "loss": 0.2484, + "step": 34640 + }, + { + "epoch": 0.5404941660947151, + "grad_norm": 0.003213417250663042, + "learning_rate": 0.00016396809550550114, + "loss": 0.1582, + "step": 34650 + }, + { + "epoch": 0.5406501528670369, + "grad_norm": 2.3433990478515625, + "learning_rate": 0.00016395769638734636, + "loss": 0.2774, + "step": 34660 + }, + { + "epoch": 0.5408061396393585, + "grad_norm": 0.09544426947832108, + "learning_rate": 0.00016394729726919158, + "loss": 0.1242, + "step": 34670 + }, + { + "epoch": 0.5409621264116803, + "grad_norm": 1.8934235572814941, + "learning_rate": 0.0001639368981510368, + "loss": 0.1636, + "step": 34680 + }, + { + "epoch": 0.541118113184002, + "grad_norm": 2.8528554439544678, + "learning_rate": 0.00016392649903288202, + "loss": 0.2524, + "step": 34690 + }, + { + "epoch": 0.5412740999563237, + "grad_norm": 2.837836980819702, + "learning_rate": 0.00016391609991472722, + "loss": 0.1808, + "step": 34700 + }, + { + "epoch": 0.5414300867286455, + "grad_norm": 4.349740028381348, + "learning_rate": 0.00016390570079657246, + "loss": 0.12, + "step": 34710 + }, + { + "epoch": 0.5415860735009671, + "grad_norm": 0.6011945605278015, + "learning_rate": 0.00016389530167841765, + "loss": 0.2865, + "step": 34720 + }, + { + "epoch": 0.5417420602732889, + "grad_norm": 1.2079488039016724, + "learning_rate": 0.0001638849025602629, + "loss": 0.125, + "step": 34730 + }, + { + "epoch": 0.5418980470456105, + "grad_norm": 1.9549164772033691, + "learning_rate": 0.0001638745034421081, + "loss": 0.3285, + "step": 34740 + }, + { + "epoch": 0.5420540338179323, + "grad_norm": 0.6969407200813293, + "learning_rate": 0.00016386410432395334, + "loss": 0.3267, + "step": 34750 + }, + { + "epoch": 0.5422100205902539, + "grad_norm": 1.0561965703964233, + "learning_rate": 0.00016385370520579853, + "loss": 0.1463, + "step": 34760 + }, + { + "epoch": 0.5423660073625757, + "grad_norm": 1.8208931684494019, + "learning_rate": 0.00016384330608764378, + "loss": 0.1713, + "step": 34770 + }, + { + "epoch": 0.5425219941348973, + "grad_norm": 0.6349910497665405, + "learning_rate": 0.000163832906969489, + "loss": 0.2244, + "step": 34780 + }, + { + "epoch": 0.5426779809072191, + "grad_norm": 1.1976515054702759, + "learning_rate": 0.00016382250785133422, + "loss": 0.3407, + "step": 34790 + }, + { + "epoch": 0.5428339676795407, + "grad_norm": 0.5970319509506226, + "learning_rate": 0.00016381210873317944, + "loss": 0.1419, + "step": 34800 + }, + { + "epoch": 0.5429899544518625, + "grad_norm": 0.14623159170150757, + "learning_rate": 0.00016380170961502466, + "loss": 0.14, + "step": 34810 + }, + { + "epoch": 0.5431459412241841, + "grad_norm": 1.4332351684570312, + "learning_rate": 0.00016379131049686988, + "loss": 0.1093, + "step": 34820 + }, + { + "epoch": 0.5433019279965059, + "grad_norm": 1.1443063020706177, + "learning_rate": 0.0001637809113787151, + "loss": 0.3676, + "step": 34830 + }, + { + "epoch": 0.5434579147688277, + "grad_norm": 1.3897461891174316, + "learning_rate": 0.00016377051226056032, + "loss": 0.3713, + "step": 34840 + }, + { + "epoch": 0.5436139015411493, + "grad_norm": 1.7773199081420898, + "learning_rate": 0.00016376011314240554, + "loss": 0.2249, + "step": 34850 + }, + { + "epoch": 0.5437698883134711, + "grad_norm": 0.8524389266967773, + "learning_rate": 0.00016374971402425076, + "loss": 0.2158, + "step": 34860 + }, + { + "epoch": 0.5439258750857927, + "grad_norm": 0.5572389960289001, + "learning_rate": 0.00016373931490609598, + "loss": 0.215, + "step": 34870 + }, + { + "epoch": 0.5440818618581145, + "grad_norm": 3.0238709449768066, + "learning_rate": 0.0001637289157879412, + "loss": 0.3765, + "step": 34880 + }, + { + "epoch": 0.5442378486304361, + "grad_norm": 1.1655421257019043, + "learning_rate": 0.00016371851666978642, + "loss": 0.2491, + "step": 34890 + }, + { + "epoch": 0.5443938354027579, + "grad_norm": 0.49492090940475464, + "learning_rate": 0.00016370811755163164, + "loss": 0.1287, + "step": 34900 + }, + { + "epoch": 0.5445498221750795, + "grad_norm": 0.2732921540737152, + "learning_rate": 0.00016369771843347685, + "loss": 0.2154, + "step": 34910 + }, + { + "epoch": 0.5447058089474013, + "grad_norm": 2.5807769298553467, + "learning_rate": 0.00016368731931532207, + "loss": 0.2032, + "step": 34920 + }, + { + "epoch": 0.5448617957197229, + "grad_norm": 0.5044315457344055, + "learning_rate": 0.0001636769201971673, + "loss": 0.2595, + "step": 34930 + }, + { + "epoch": 0.5450177824920447, + "grad_norm": 2.1248972415924072, + "learning_rate": 0.00016366652107901251, + "loss": 0.5651, + "step": 34940 + }, + { + "epoch": 0.5451737692643663, + "grad_norm": 0.8391468524932861, + "learning_rate": 0.00016365612196085773, + "loss": 0.3028, + "step": 34950 + }, + { + "epoch": 0.5453297560366881, + "grad_norm": 0.36081477999687195, + "learning_rate": 0.00016364572284270295, + "loss": 0.1073, + "step": 34960 + }, + { + "epoch": 0.5454857428090097, + "grad_norm": 3.175804376602173, + "learning_rate": 0.00016363532372454817, + "loss": 0.1889, + "step": 34970 + }, + { + "epoch": 0.5456417295813315, + "grad_norm": 0.20647937059402466, + "learning_rate": 0.0001636249246063934, + "loss": 0.2416, + "step": 34980 + }, + { + "epoch": 0.5457977163536533, + "grad_norm": 2.3525943756103516, + "learning_rate": 0.0001636145254882386, + "loss": 0.3179, + "step": 34990 + }, + { + "epoch": 0.5459537031259749, + "grad_norm": 0.2641216814517975, + "learning_rate": 0.00016360412637008383, + "loss": 0.2157, + "step": 35000 + }, + { + "epoch": 0.5461096898982967, + "grad_norm": 0.09906073659658432, + "learning_rate": 0.00016359372725192905, + "loss": 0.2576, + "step": 35010 + }, + { + "epoch": 0.5462656766706183, + "grad_norm": 0.13971920311450958, + "learning_rate": 0.00016358332813377427, + "loss": 0.1298, + "step": 35020 + }, + { + "epoch": 0.5464216634429401, + "grad_norm": 1.4535638093948364, + "learning_rate": 0.0001635729290156195, + "loss": 0.2357, + "step": 35030 + }, + { + "epoch": 0.5465776502152617, + "grad_norm": 0.9107828736305237, + "learning_rate": 0.0001635625298974647, + "loss": 0.3049, + "step": 35040 + }, + { + "epoch": 0.5467336369875835, + "grad_norm": 0.0377386212348938, + "learning_rate": 0.00016355213077930993, + "loss": 0.2066, + "step": 35050 + }, + { + "epoch": 0.5468896237599051, + "grad_norm": 2.7800869941711426, + "learning_rate": 0.00016354173166115515, + "loss": 0.2974, + "step": 35060 + }, + { + "epoch": 0.5470456105322269, + "grad_norm": 1.6247998476028442, + "learning_rate": 0.00016353133254300037, + "loss": 0.3822, + "step": 35070 + }, + { + "epoch": 0.5472015973045485, + "grad_norm": 1.6479015350341797, + "learning_rate": 0.0001635209334248456, + "loss": 0.3401, + "step": 35080 + }, + { + "epoch": 0.5473575840768703, + "grad_norm": 0.1879737675189972, + "learning_rate": 0.0001635105343066908, + "loss": 0.1549, + "step": 35090 + }, + { + "epoch": 0.5475135708491919, + "grad_norm": 0.39355361461639404, + "learning_rate": 0.00016350013518853603, + "loss": 0.2032, + "step": 35100 + }, + { + "epoch": 0.5476695576215137, + "grad_norm": 2.9976983070373535, + "learning_rate": 0.00016348973607038125, + "loss": 0.3867, + "step": 35110 + }, + { + "epoch": 0.5478255443938354, + "grad_norm": 0.8242707252502441, + "learning_rate": 0.00016347933695222647, + "loss": 0.3601, + "step": 35120 + }, + { + "epoch": 0.5479815311661571, + "grad_norm": 0.874955415725708, + "learning_rate": 0.0001634689378340717, + "loss": 0.2709, + "step": 35130 + }, + { + "epoch": 0.5481375179384789, + "grad_norm": 0.05980971083045006, + "learning_rate": 0.0001634585387159169, + "loss": 0.2646, + "step": 35140 + }, + { + "epoch": 0.5482935047108005, + "grad_norm": 1.8437328338623047, + "learning_rate": 0.0001634481395977621, + "loss": 0.1965, + "step": 35150 + }, + { + "epoch": 0.5484494914831223, + "grad_norm": 0.4142405688762665, + "learning_rate": 0.00016343774047960735, + "loss": 0.1756, + "step": 35160 + }, + { + "epoch": 0.5486054782554439, + "grad_norm": 4.836324214935303, + "learning_rate": 0.00016342734136145254, + "loss": 0.2534, + "step": 35170 + }, + { + "epoch": 0.5487614650277657, + "grad_norm": 1.4933065176010132, + "learning_rate": 0.00016341694224329779, + "loss": 0.2355, + "step": 35180 + }, + { + "epoch": 0.5489174518000873, + "grad_norm": 1.9468894004821777, + "learning_rate": 0.00016340654312514298, + "loss": 0.3113, + "step": 35190 + }, + { + "epoch": 0.5490734385724091, + "grad_norm": 1.012710452079773, + "learning_rate": 0.00016339614400698822, + "loss": 0.2635, + "step": 35200 + }, + { + "epoch": 0.5492294253447307, + "grad_norm": 1.5551140308380127, + "learning_rate": 0.00016338574488883342, + "loss": 0.2951, + "step": 35210 + }, + { + "epoch": 0.5493854121170525, + "grad_norm": 1.8613696098327637, + "learning_rate": 0.00016337534577067866, + "loss": 0.2832, + "step": 35220 + }, + { + "epoch": 0.5495413988893741, + "grad_norm": 1.6748839616775513, + "learning_rate": 0.00016336494665252386, + "loss": 0.2266, + "step": 35230 + }, + { + "epoch": 0.5496973856616959, + "grad_norm": 0.4094032347202301, + "learning_rate": 0.0001633545475343691, + "loss": 0.3497, + "step": 35240 + }, + { + "epoch": 0.5498533724340176, + "grad_norm": 1.6414631605148315, + "learning_rate": 0.0001633441484162143, + "loss": 0.1763, + "step": 35250 + }, + { + "epoch": 0.5500093592063393, + "grad_norm": 1.6380645036697388, + "learning_rate": 0.00016333374929805954, + "loss": 0.1355, + "step": 35260 + }, + { + "epoch": 0.550165345978661, + "grad_norm": 0.7325630187988281, + "learning_rate": 0.00016332335017990473, + "loss": 0.3105, + "step": 35270 + }, + { + "epoch": 0.5503213327509827, + "grad_norm": 1.6015644073486328, + "learning_rate": 0.00016331295106174998, + "loss": 0.2311, + "step": 35280 + }, + { + "epoch": 0.5504773195233045, + "grad_norm": 0.9224210381507874, + "learning_rate": 0.00016330255194359517, + "loss": 0.2746, + "step": 35290 + }, + { + "epoch": 0.5506333062956261, + "grad_norm": 1.30025315284729, + "learning_rate": 0.00016329215282544042, + "loss": 0.2066, + "step": 35300 + }, + { + "epoch": 0.5507892930679479, + "grad_norm": 0.481125146150589, + "learning_rate": 0.0001632817537072856, + "loss": 0.1778, + "step": 35310 + }, + { + "epoch": 0.5509452798402695, + "grad_norm": 0.2709486782550812, + "learning_rate": 0.00016327135458913086, + "loss": 0.1131, + "step": 35320 + }, + { + "epoch": 0.5511012666125913, + "grad_norm": 0.6248563528060913, + "learning_rate": 0.00016326095547097605, + "loss": 0.4356, + "step": 35330 + }, + { + "epoch": 0.5512572533849129, + "grad_norm": 3.0947647094726562, + "learning_rate": 0.0001632505563528213, + "loss": 0.3072, + "step": 35340 + }, + { + "epoch": 0.5514132401572347, + "grad_norm": 1.009535312652588, + "learning_rate": 0.0001632401572346665, + "loss": 0.1808, + "step": 35350 + }, + { + "epoch": 0.5515692269295563, + "grad_norm": 0.042604975402355194, + "learning_rate": 0.00016322975811651174, + "loss": 0.1198, + "step": 35360 + }, + { + "epoch": 0.5517252137018781, + "grad_norm": 1.8652396202087402, + "learning_rate": 0.00016321935899835693, + "loss": 0.2176, + "step": 35370 + }, + { + "epoch": 0.5518812004741998, + "grad_norm": 4.6097187995910645, + "learning_rate": 0.00016320895988020218, + "loss": 0.2132, + "step": 35380 + }, + { + "epoch": 0.5520371872465215, + "grad_norm": 1.3615522384643555, + "learning_rate": 0.00016319856076204737, + "loss": 0.2751, + "step": 35390 + }, + { + "epoch": 0.5521931740188432, + "grad_norm": 0.8844773769378662, + "learning_rate": 0.00016318816164389262, + "loss": 0.22, + "step": 35400 + }, + { + "epoch": 0.5523491607911649, + "grad_norm": 0.5413331985473633, + "learning_rate": 0.0001631777625257378, + "loss": 0.2159, + "step": 35410 + }, + { + "epoch": 0.5525051475634866, + "grad_norm": 1.5456678867340088, + "learning_rate": 0.00016316736340758306, + "loss": 0.2244, + "step": 35420 + }, + { + "epoch": 0.5526611343358083, + "grad_norm": 2.1405861377716064, + "learning_rate": 0.00016315696428942825, + "loss": 0.446, + "step": 35430 + }, + { + "epoch": 0.5528171211081301, + "grad_norm": 1.4269858598709106, + "learning_rate": 0.0001631465651712735, + "loss": 0.3599, + "step": 35440 + }, + { + "epoch": 0.5529731078804517, + "grad_norm": 1.8682516813278198, + "learning_rate": 0.0001631361660531187, + "loss": 0.115, + "step": 35450 + }, + { + "epoch": 0.5531290946527735, + "grad_norm": 1.0175774097442627, + "learning_rate": 0.00016312576693496394, + "loss": 0.4003, + "step": 35460 + }, + { + "epoch": 0.5532850814250951, + "grad_norm": 2.4369170665740967, + "learning_rate": 0.00016311536781680913, + "loss": 0.2489, + "step": 35470 + }, + { + "epoch": 0.5534410681974169, + "grad_norm": 1.886022686958313, + "learning_rate": 0.00016310496869865437, + "loss": 0.4201, + "step": 35480 + }, + { + "epoch": 0.5535970549697385, + "grad_norm": 1.5811959505081177, + "learning_rate": 0.00016309456958049957, + "loss": 0.4286, + "step": 35490 + }, + { + "epoch": 0.5537530417420603, + "grad_norm": 1.971110224723816, + "learning_rate": 0.00016308417046234481, + "loss": 0.3599, + "step": 35500 + }, + { + "epoch": 0.553909028514382, + "grad_norm": 0.0335380844771862, + "learning_rate": 0.00016307377134419, + "loss": 0.3357, + "step": 35510 + }, + { + "epoch": 0.5540650152867037, + "grad_norm": 1.8519577980041504, + "learning_rate": 0.00016306337222603525, + "loss": 0.3126, + "step": 35520 + }, + { + "epoch": 0.5542210020590254, + "grad_norm": 2.1563379764556885, + "learning_rate": 0.00016305297310788045, + "loss": 0.1123, + "step": 35530 + }, + { + "epoch": 0.5543769888313471, + "grad_norm": 1.64332914352417, + "learning_rate": 0.0001630425739897257, + "loss": 0.2853, + "step": 35540 + }, + { + "epoch": 0.5545329756036688, + "grad_norm": 0.061150554567575455, + "learning_rate": 0.00016303217487157088, + "loss": 0.1943, + "step": 35550 + }, + { + "epoch": 0.5546889623759905, + "grad_norm": 1.2701060771942139, + "learning_rate": 0.00016302177575341613, + "loss": 0.2304, + "step": 35560 + }, + { + "epoch": 0.5548449491483122, + "grad_norm": 2.424860715866089, + "learning_rate": 0.00016301137663526132, + "loss": 0.2128, + "step": 35570 + }, + { + "epoch": 0.5550009359206339, + "grad_norm": 0.6803575158119202, + "learning_rate": 0.00016300097751710657, + "loss": 0.1145, + "step": 35580 + }, + { + "epoch": 0.5551569226929557, + "grad_norm": 1.2855092287063599, + "learning_rate": 0.00016299057839895176, + "loss": 0.2448, + "step": 35590 + }, + { + "epoch": 0.5553129094652773, + "grad_norm": 2.6340911388397217, + "learning_rate": 0.00016298017928079698, + "loss": 0.1856, + "step": 35600 + }, + { + "epoch": 0.5554688962375991, + "grad_norm": 1.776382327079773, + "learning_rate": 0.0001629697801626422, + "loss": 0.2422, + "step": 35610 + }, + { + "epoch": 0.5556248830099207, + "grad_norm": 3.0746347904205322, + "learning_rate": 0.00016295938104448742, + "loss": 0.1578, + "step": 35620 + }, + { + "epoch": 0.5557808697822425, + "grad_norm": 1.4887659549713135, + "learning_rate": 0.00016294898192633267, + "loss": 0.2385, + "step": 35630 + }, + { + "epoch": 0.5559368565545642, + "grad_norm": 0.5371220707893372, + "learning_rate": 0.00016293858280817786, + "loss": 0.2361, + "step": 35640 + }, + { + "epoch": 0.5560928433268859, + "grad_norm": 0.18604904413223267, + "learning_rate": 0.0001629281836900231, + "loss": 0.0828, + "step": 35650 + }, + { + "epoch": 0.5562488300992076, + "grad_norm": 2.267854690551758, + "learning_rate": 0.0001629177845718683, + "loss": 0.413, + "step": 35660 + }, + { + "epoch": 0.5564048168715293, + "grad_norm": 1.9755452871322632, + "learning_rate": 0.00016290738545371355, + "loss": 0.146, + "step": 35670 + }, + { + "epoch": 0.556560803643851, + "grad_norm": 2.430293321609497, + "learning_rate": 0.00016289698633555874, + "loss": 0.2677, + "step": 35680 + }, + { + "epoch": 0.5567167904161727, + "grad_norm": 0.6319543719291687, + "learning_rate": 0.000162886587217404, + "loss": 0.1481, + "step": 35690 + }, + { + "epoch": 0.5568727771884944, + "grad_norm": 0.17439277470111847, + "learning_rate": 0.00016287618809924918, + "loss": 0.1641, + "step": 35700 + }, + { + "epoch": 0.5570287639608161, + "grad_norm": 0.350175142288208, + "learning_rate": 0.00016286578898109443, + "loss": 0.393, + "step": 35710 + }, + { + "epoch": 0.5571847507331378, + "grad_norm": 2.641941547393799, + "learning_rate": 0.00016285538986293962, + "loss": 0.1907, + "step": 35720 + }, + { + "epoch": 0.5573407375054595, + "grad_norm": 0.1910303384065628, + "learning_rate": 0.00016284499074478487, + "loss": 0.2248, + "step": 35730 + }, + { + "epoch": 0.5574967242777813, + "grad_norm": 1.0180896520614624, + "learning_rate": 0.00016283459162663006, + "loss": 0.1924, + "step": 35740 + }, + { + "epoch": 0.557652711050103, + "grad_norm": 1.2047260999679565, + "learning_rate": 0.0001628241925084753, + "loss": 0.2031, + "step": 35750 + }, + { + "epoch": 0.5578086978224247, + "grad_norm": 1.8182405233383179, + "learning_rate": 0.0001628137933903205, + "loss": 0.1573, + "step": 35760 + }, + { + "epoch": 0.5579646845947464, + "grad_norm": 0.5485963821411133, + "learning_rate": 0.00016280339427216574, + "loss": 0.152, + "step": 35770 + }, + { + "epoch": 0.5581206713670681, + "grad_norm": 0.1674145758152008, + "learning_rate": 0.00016279299515401094, + "loss": 0.2388, + "step": 35780 + }, + { + "epoch": 0.5582766581393898, + "grad_norm": 0.36187657713890076, + "learning_rate": 0.00016278259603585618, + "loss": 0.1618, + "step": 35790 + }, + { + "epoch": 0.5584326449117115, + "grad_norm": 0.013768521137535572, + "learning_rate": 0.00016277219691770138, + "loss": 0.3666, + "step": 35800 + }, + { + "epoch": 0.5585886316840332, + "grad_norm": 2.8602802753448486, + "learning_rate": 0.00016276179779954662, + "loss": 0.2065, + "step": 35810 + }, + { + "epoch": 0.5587446184563549, + "grad_norm": 1.7925455570220947, + "learning_rate": 0.00016275139868139182, + "loss": 0.1729, + "step": 35820 + }, + { + "epoch": 0.5589006052286766, + "grad_norm": 2.910456418991089, + "learning_rate": 0.00016274099956323706, + "loss": 0.3734, + "step": 35830 + }, + { + "epoch": 0.5590565920009983, + "grad_norm": 0.8378308415412903, + "learning_rate": 0.00016273060044508225, + "loss": 0.2255, + "step": 35840 + }, + { + "epoch": 0.55921257877332, + "grad_norm": 0.409534752368927, + "learning_rate": 0.0001627202013269275, + "loss": 0.2046, + "step": 35850 + }, + { + "epoch": 0.5593685655456417, + "grad_norm": 0.49498099088668823, + "learning_rate": 0.0001627098022087727, + "loss": 0.1058, + "step": 35860 + }, + { + "epoch": 0.5595245523179634, + "grad_norm": 1.1761783361434937, + "learning_rate": 0.00016269940309061794, + "loss": 0.1566, + "step": 35870 + }, + { + "epoch": 0.5596805390902851, + "grad_norm": 3.8204751014709473, + "learning_rate": 0.00016268900397246313, + "loss": 0.2646, + "step": 35880 + }, + { + "epoch": 0.5598365258626069, + "grad_norm": 0.9882522225379944, + "learning_rate": 0.00016267860485430838, + "loss": 0.1756, + "step": 35890 + }, + { + "epoch": 0.5599925126349286, + "grad_norm": 1.1832259893417358, + "learning_rate": 0.00016266820573615357, + "loss": 0.1385, + "step": 35900 + }, + { + "epoch": 0.5601484994072503, + "grad_norm": 0.7638296484947205, + "learning_rate": 0.00016265780661799882, + "loss": 0.5679, + "step": 35910 + }, + { + "epoch": 0.560304486179572, + "grad_norm": 0.8551504611968994, + "learning_rate": 0.000162647407499844, + "loss": 0.2981, + "step": 35920 + }, + { + "epoch": 0.5604604729518937, + "grad_norm": 5.499948501586914, + "learning_rate": 0.00016263700838168926, + "loss": 0.1429, + "step": 35930 + }, + { + "epoch": 0.5606164597242154, + "grad_norm": 3.0494496822357178, + "learning_rate": 0.00016262660926353445, + "loss": 0.1322, + "step": 35940 + }, + { + "epoch": 0.5607724464965371, + "grad_norm": 0.2482384592294693, + "learning_rate": 0.0001626162101453797, + "loss": 0.2473, + "step": 35950 + }, + { + "epoch": 0.5609284332688588, + "grad_norm": 0.009750776924192905, + "learning_rate": 0.0001626058110272249, + "loss": 0.0429, + "step": 35960 + }, + { + "epoch": 0.5610844200411805, + "grad_norm": 0.5742604732513428, + "learning_rate": 0.00016259541190907014, + "loss": 0.4565, + "step": 35970 + }, + { + "epoch": 0.5612404068135022, + "grad_norm": 0.7810243368148804, + "learning_rate": 0.00016258501279091533, + "loss": 0.2107, + "step": 35980 + }, + { + "epoch": 0.5613963935858239, + "grad_norm": 2.785747528076172, + "learning_rate": 0.00016257461367276058, + "loss": 0.5087, + "step": 35990 + }, + { + "epoch": 0.5615523803581456, + "grad_norm": 0.8806902766227722, + "learning_rate": 0.00016256421455460577, + "loss": 0.2937, + "step": 36000 + }, + { + "epoch": 0.5617083671304673, + "grad_norm": 1.857373833656311, + "learning_rate": 0.00016255381543645102, + "loss": 0.5042, + "step": 36010 + }, + { + "epoch": 0.561864353902789, + "grad_norm": 0.5333901047706604, + "learning_rate": 0.0001625434163182962, + "loss": 0.1678, + "step": 36020 + }, + { + "epoch": 0.5620203406751108, + "grad_norm": 11.023160934448242, + "learning_rate": 0.00016253301720014145, + "loss": 0.2721, + "step": 36030 + }, + { + "epoch": 0.5621763274474325, + "grad_norm": 2.6214029788970947, + "learning_rate": 0.00016252261808198665, + "loss": 0.1675, + "step": 36040 + }, + { + "epoch": 0.5623323142197542, + "grad_norm": 0.25980064272880554, + "learning_rate": 0.0001625122189638319, + "loss": 0.1832, + "step": 36050 + }, + { + "epoch": 0.5624883009920759, + "grad_norm": 1.3559473752975464, + "learning_rate": 0.0001625018198456771, + "loss": 0.121, + "step": 36060 + }, + { + "epoch": 0.5626442877643976, + "grad_norm": 0.3558153212070465, + "learning_rate": 0.0001624914207275223, + "loss": 0.2797, + "step": 36070 + }, + { + "epoch": 0.5628002745367193, + "grad_norm": 0.6314427256584167, + "learning_rate": 0.00016248102160936753, + "loss": 0.3455, + "step": 36080 + }, + { + "epoch": 0.562956261309041, + "grad_norm": 0.16120101511478424, + "learning_rate": 0.00016247062249121275, + "loss": 0.3669, + "step": 36090 + }, + { + "epoch": 0.5631122480813627, + "grad_norm": 1.041572093963623, + "learning_rate": 0.00016246022337305797, + "loss": 0.2744, + "step": 36100 + }, + { + "epoch": 0.5632682348536844, + "grad_norm": 0.2645890414714813, + "learning_rate": 0.00016244982425490318, + "loss": 0.1487, + "step": 36110 + }, + { + "epoch": 0.5634242216260061, + "grad_norm": 5.698698043823242, + "learning_rate": 0.0001624394251367484, + "loss": 0.173, + "step": 36120 + }, + { + "epoch": 0.5635802083983278, + "grad_norm": 0.3299804627895355, + "learning_rate": 0.00016242902601859362, + "loss": 0.1965, + "step": 36130 + }, + { + "epoch": 0.5637361951706495, + "grad_norm": 1.2476481199264526, + "learning_rate": 0.00016241862690043884, + "loss": 0.4097, + "step": 36140 + }, + { + "epoch": 0.5638921819429712, + "grad_norm": 0.30926263332366943, + "learning_rate": 0.00016240822778228406, + "loss": 0.1416, + "step": 36150 + }, + { + "epoch": 0.564048168715293, + "grad_norm": 0.05872740224003792, + "learning_rate": 0.00016239782866412928, + "loss": 0.1449, + "step": 36160 + }, + { + "epoch": 0.5642041554876146, + "grad_norm": 0.2292211800813675, + "learning_rate": 0.0001623874295459745, + "loss": 0.2557, + "step": 36170 + }, + { + "epoch": 0.5643601422599364, + "grad_norm": 1.7822531461715698, + "learning_rate": 0.00016237703042781972, + "loss": 0.2953, + "step": 36180 + }, + { + "epoch": 0.5645161290322581, + "grad_norm": 2.4908461570739746, + "learning_rate": 0.00016236663130966494, + "loss": 0.3888, + "step": 36190 + }, + { + "epoch": 0.5646721158045798, + "grad_norm": 0.08363594114780426, + "learning_rate": 0.00016235623219151016, + "loss": 0.2293, + "step": 36200 + }, + { + "epoch": 0.5648281025769015, + "grad_norm": 1.4830002784729004, + "learning_rate": 0.00016234583307335538, + "loss": 0.2348, + "step": 36210 + }, + { + "epoch": 0.5649840893492232, + "grad_norm": 1.8443071842193604, + "learning_rate": 0.0001623354339552006, + "loss": 0.3392, + "step": 36220 + }, + { + "epoch": 0.5651400761215449, + "grad_norm": 2.8050875663757324, + "learning_rate": 0.00016232503483704582, + "loss": 0.2426, + "step": 36230 + }, + { + "epoch": 0.5652960628938666, + "grad_norm": 3.6627259254455566, + "learning_rate": 0.00016231463571889104, + "loss": 0.2118, + "step": 36240 + }, + { + "epoch": 0.5654520496661883, + "grad_norm": 0.3443094789981842, + "learning_rate": 0.00016230423660073626, + "loss": 0.2418, + "step": 36250 + }, + { + "epoch": 0.56560803643851, + "grad_norm": 1.7411112785339355, + "learning_rate": 0.00016229383748258148, + "loss": 0.1628, + "step": 36260 + }, + { + "epoch": 0.5657640232108317, + "grad_norm": 1.1321816444396973, + "learning_rate": 0.0001622834383644267, + "loss": 0.2732, + "step": 36270 + }, + { + "epoch": 0.5659200099831534, + "grad_norm": 2.9686615467071533, + "learning_rate": 0.00016227303924627192, + "loss": 0.3036, + "step": 36280 + }, + { + "epoch": 0.5660759967554752, + "grad_norm": 5.03530740737915, + "learning_rate": 0.00016226264012811714, + "loss": 0.2143, + "step": 36290 + }, + { + "epoch": 0.5662319835277968, + "grad_norm": 2.3943281173706055, + "learning_rate": 0.00016225224100996236, + "loss": 0.2629, + "step": 36300 + }, + { + "epoch": 0.5663879703001186, + "grad_norm": 1.812828779220581, + "learning_rate": 0.00016224184189180758, + "loss": 0.3892, + "step": 36310 + }, + { + "epoch": 0.5665439570724402, + "grad_norm": 1.4708483219146729, + "learning_rate": 0.0001622314427736528, + "loss": 0.2118, + "step": 36320 + }, + { + "epoch": 0.566699943844762, + "grad_norm": 8.99713134765625, + "learning_rate": 0.00016222104365549802, + "loss": 0.3185, + "step": 36330 + }, + { + "epoch": 0.5668559306170837, + "grad_norm": 1.7472341060638428, + "learning_rate": 0.00016221064453734324, + "loss": 0.1659, + "step": 36340 + }, + { + "epoch": 0.5670119173894054, + "grad_norm": 0.8059778213500977, + "learning_rate": 0.00016220024541918846, + "loss": 0.0621, + "step": 36350 + }, + { + "epoch": 0.5671679041617271, + "grad_norm": 4.0501861572265625, + "learning_rate": 0.00016218984630103368, + "loss": 0.1399, + "step": 36360 + }, + { + "epoch": 0.5673238909340488, + "grad_norm": 0.014400321058928967, + "learning_rate": 0.0001621794471828789, + "loss": 0.2372, + "step": 36370 + }, + { + "epoch": 0.5674798777063705, + "grad_norm": 0.4061933755874634, + "learning_rate": 0.00016216904806472412, + "loss": 0.3854, + "step": 36380 + }, + { + "epoch": 0.5676358644786922, + "grad_norm": 0.4940955340862274, + "learning_rate": 0.00016215864894656933, + "loss": 0.1689, + "step": 36390 + }, + { + "epoch": 0.567791851251014, + "grad_norm": 0.3220517635345459, + "learning_rate": 0.00016214824982841455, + "loss": 0.1912, + "step": 36400 + }, + { + "epoch": 0.5679478380233356, + "grad_norm": 2.0848920345306396, + "learning_rate": 0.00016213785071025977, + "loss": 0.1639, + "step": 36410 + }, + { + "epoch": 0.5681038247956574, + "grad_norm": 0.04129406809806824, + "learning_rate": 0.000162127451592105, + "loss": 0.1378, + "step": 36420 + }, + { + "epoch": 0.568259811567979, + "grad_norm": 1.9343401193618774, + "learning_rate": 0.0001621170524739502, + "loss": 0.3759, + "step": 36430 + }, + { + "epoch": 0.5684157983403008, + "grad_norm": 1.0625615119934082, + "learning_rate": 0.00016210665335579543, + "loss": 0.2074, + "step": 36440 + }, + { + "epoch": 0.5685717851126224, + "grad_norm": 0.14784491062164307, + "learning_rate": 0.00016209625423764065, + "loss": 0.0981, + "step": 36450 + }, + { + "epoch": 0.5687277718849442, + "grad_norm": 1.8117821216583252, + "learning_rate": 0.00016208585511948587, + "loss": 0.2755, + "step": 36460 + }, + { + "epoch": 0.5688837586572658, + "grad_norm": 0.061010006815195084, + "learning_rate": 0.0001620754560013311, + "loss": 0.1815, + "step": 36470 + }, + { + "epoch": 0.5690397454295876, + "grad_norm": 0.44453224539756775, + "learning_rate": 0.0001620650568831763, + "loss": 0.0891, + "step": 36480 + }, + { + "epoch": 0.5691957322019093, + "grad_norm": 0.6130645871162415, + "learning_rate": 0.00016205465776502153, + "loss": 0.2778, + "step": 36490 + }, + { + "epoch": 0.569351718974231, + "grad_norm": 5.408796787261963, + "learning_rate": 0.00016204425864686678, + "loss": 0.2301, + "step": 36500 + }, + { + "epoch": 0.5695077057465527, + "grad_norm": 1.2095412015914917, + "learning_rate": 0.00016203385952871197, + "loss": 0.1378, + "step": 36510 + }, + { + "epoch": 0.5696636925188744, + "grad_norm": 0.46873530745506287, + "learning_rate": 0.0001620234604105572, + "loss": 0.1431, + "step": 36520 + }, + { + "epoch": 0.5698196792911961, + "grad_norm": 1.028438925743103, + "learning_rate": 0.0001620130612924024, + "loss": 0.1114, + "step": 36530 + }, + { + "epoch": 0.5699756660635178, + "grad_norm": 0.5952587127685547, + "learning_rate": 0.00016200266217424763, + "loss": 0.2156, + "step": 36540 + }, + { + "epoch": 0.5701316528358396, + "grad_norm": 0.5134285688400269, + "learning_rate": 0.00016199226305609285, + "loss": 0.2712, + "step": 36550 + }, + { + "epoch": 0.5702876396081612, + "grad_norm": 0.03440163657069206, + "learning_rate": 0.00016198186393793807, + "loss": 0.312, + "step": 36560 + }, + { + "epoch": 0.570443626380483, + "grad_norm": 0.10363951325416565, + "learning_rate": 0.0001619714648197833, + "loss": 0.2399, + "step": 36570 + }, + { + "epoch": 0.5705996131528046, + "grad_norm": 1.3341947793960571, + "learning_rate": 0.0001619610657016285, + "loss": 0.1397, + "step": 36580 + }, + { + "epoch": 0.5707555999251264, + "grad_norm": 0.0668073296546936, + "learning_rate": 0.00016195066658347373, + "loss": 0.1091, + "step": 36590 + }, + { + "epoch": 0.570911586697448, + "grad_norm": 5.0486674308776855, + "learning_rate": 0.00016194026746531895, + "loss": 0.4893, + "step": 36600 + }, + { + "epoch": 0.5710675734697698, + "grad_norm": 1.2655692100524902, + "learning_rate": 0.00016192986834716417, + "loss": 0.3344, + "step": 36610 + }, + { + "epoch": 0.5712235602420914, + "grad_norm": 0.1458190381526947, + "learning_rate": 0.00016191946922900939, + "loss": 0.247, + "step": 36620 + }, + { + "epoch": 0.5713795470144132, + "grad_norm": 2.5789072513580322, + "learning_rate": 0.0001619090701108546, + "loss": 0.1546, + "step": 36630 + }, + { + "epoch": 0.5715355337867349, + "grad_norm": 1.191821575164795, + "learning_rate": 0.00016189867099269983, + "loss": 0.2215, + "step": 36640 + }, + { + "epoch": 0.5716915205590566, + "grad_norm": 8.163617134094238, + "learning_rate": 0.00016188827187454505, + "loss": 0.1696, + "step": 36650 + }, + { + "epoch": 0.5718475073313783, + "grad_norm": 3.162277936935425, + "learning_rate": 0.00016187787275639027, + "loss": 0.2906, + "step": 36660 + }, + { + "epoch": 0.5720034941037, + "grad_norm": 2.2368602752685547, + "learning_rate": 0.00016186747363823548, + "loss": 0.3806, + "step": 36670 + }, + { + "epoch": 0.5721594808760218, + "grad_norm": 2.0193803310394287, + "learning_rate": 0.0001618570745200807, + "loss": 0.1964, + "step": 36680 + }, + { + "epoch": 0.5723154676483434, + "grad_norm": 0.9131811261177063, + "learning_rate": 0.00016184667540192592, + "loss": 0.1902, + "step": 36690 + }, + { + "epoch": 0.5724714544206652, + "grad_norm": 0.21330870687961578, + "learning_rate": 0.00016183627628377114, + "loss": 0.2907, + "step": 36700 + }, + { + "epoch": 0.5726274411929868, + "grad_norm": 0.7145895957946777, + "learning_rate": 0.00016182587716561636, + "loss": 0.2574, + "step": 36710 + }, + { + "epoch": 0.5727834279653086, + "grad_norm": 1.0699433088302612, + "learning_rate": 0.00016181547804746158, + "loss": 0.1675, + "step": 36720 + }, + { + "epoch": 0.5729394147376302, + "grad_norm": 0.5066679120063782, + "learning_rate": 0.0001618050789293068, + "loss": 0.4487, + "step": 36730 + }, + { + "epoch": 0.573095401509952, + "grad_norm": 0.7482333183288574, + "learning_rate": 0.00016179467981115202, + "loss": 0.4101, + "step": 36740 + }, + { + "epoch": 0.5732513882822736, + "grad_norm": 0.15925532579421997, + "learning_rate": 0.00016178428069299724, + "loss": 0.2825, + "step": 36750 + }, + { + "epoch": 0.5734073750545954, + "grad_norm": 0.95689457654953, + "learning_rate": 0.00016177388157484246, + "loss": 0.2754, + "step": 36760 + }, + { + "epoch": 0.573563361826917, + "grad_norm": 2.15043568611145, + "learning_rate": 0.00016176348245668768, + "loss": 0.2073, + "step": 36770 + }, + { + "epoch": 0.5737193485992388, + "grad_norm": 2.651520252227783, + "learning_rate": 0.0001617530833385329, + "loss": 0.2517, + "step": 36780 + }, + { + "epoch": 0.5738753353715605, + "grad_norm": 2.1156702041625977, + "learning_rate": 0.00016174268422037812, + "loss": 0.3029, + "step": 36790 + }, + { + "epoch": 0.5740313221438822, + "grad_norm": 3.6200709342956543, + "learning_rate": 0.00016173228510222334, + "loss": 0.4174, + "step": 36800 + }, + { + "epoch": 0.574187308916204, + "grad_norm": 1.447937250137329, + "learning_rate": 0.00016172188598406856, + "loss": 0.1763, + "step": 36810 + }, + { + "epoch": 0.5743432956885256, + "grad_norm": 1.4020819664001465, + "learning_rate": 0.00016171148686591378, + "loss": 0.3994, + "step": 36820 + }, + { + "epoch": 0.5744992824608474, + "grad_norm": 1.4079807996749878, + "learning_rate": 0.000161701087747759, + "loss": 0.2051, + "step": 36830 + }, + { + "epoch": 0.574655269233169, + "grad_norm": 0.05268567427992821, + "learning_rate": 0.00016169068862960422, + "loss": 0.1773, + "step": 36840 + }, + { + "epoch": 0.5748112560054908, + "grad_norm": 2.248767614364624, + "learning_rate": 0.00016168028951144944, + "loss": 0.2534, + "step": 36850 + }, + { + "epoch": 0.5749672427778124, + "grad_norm": 0.3726302981376648, + "learning_rate": 0.00016166989039329466, + "loss": 0.328, + "step": 36860 + }, + { + "epoch": 0.5751232295501342, + "grad_norm": 1.0772110223770142, + "learning_rate": 0.00016165949127513988, + "loss": 0.2069, + "step": 36870 + }, + { + "epoch": 0.5752792163224558, + "grad_norm": 0.11339398473501205, + "learning_rate": 0.0001616490921569851, + "loss": 0.1827, + "step": 36880 + }, + { + "epoch": 0.5754352030947776, + "grad_norm": 1.2981103658676147, + "learning_rate": 0.00016163869303883032, + "loss": 0.1378, + "step": 36890 + }, + { + "epoch": 0.5755911898670992, + "grad_norm": 0.3302285075187683, + "learning_rate": 0.00016162829392067554, + "loss": 0.2311, + "step": 36900 + }, + { + "epoch": 0.575747176639421, + "grad_norm": 0.10168436169624329, + "learning_rate": 0.00016161789480252076, + "loss": 0.1678, + "step": 36910 + }, + { + "epoch": 0.5759031634117426, + "grad_norm": 1.6778357028961182, + "learning_rate": 0.00016160749568436598, + "loss": 0.2385, + "step": 36920 + }, + { + "epoch": 0.5760591501840644, + "grad_norm": 8.1710205078125, + "learning_rate": 0.0001615970965662112, + "loss": 0.3287, + "step": 36930 + }, + { + "epoch": 0.5762151369563862, + "grad_norm": 2.6481053829193115, + "learning_rate": 0.00016158669744805641, + "loss": 0.3119, + "step": 36940 + }, + { + "epoch": 0.5763711237287078, + "grad_norm": 0.3120105266571045, + "learning_rate": 0.00016157629832990163, + "loss": 0.1985, + "step": 36950 + }, + { + "epoch": 0.5765271105010296, + "grad_norm": 2.7221410274505615, + "learning_rate": 0.00016156589921174685, + "loss": 0.3697, + "step": 36960 + }, + { + "epoch": 0.5766830972733512, + "grad_norm": 0.34397706389427185, + "learning_rate": 0.00016155550009359207, + "loss": 0.2162, + "step": 36970 + }, + { + "epoch": 0.576839084045673, + "grad_norm": 4.6421003341674805, + "learning_rate": 0.0001615451009754373, + "loss": 0.2892, + "step": 36980 + }, + { + "epoch": 0.5769950708179946, + "grad_norm": 1.5395855903625488, + "learning_rate": 0.0001615347018572825, + "loss": 0.1258, + "step": 36990 + }, + { + "epoch": 0.5771510575903164, + "grad_norm": 0.4586782455444336, + "learning_rate": 0.00016152430273912773, + "loss": 0.1752, + "step": 37000 + }, + { + "epoch": 0.577307044362638, + "grad_norm": 0.4277658760547638, + "learning_rate": 0.00016151390362097295, + "loss": 0.1722, + "step": 37010 + }, + { + "epoch": 0.5774630311349598, + "grad_norm": 1.0282249450683594, + "learning_rate": 0.00016150350450281817, + "loss": 0.1361, + "step": 37020 + }, + { + "epoch": 0.5776190179072814, + "grad_norm": 0.2496921867132187, + "learning_rate": 0.0001614931053846634, + "loss": 0.0954, + "step": 37030 + }, + { + "epoch": 0.5777750046796032, + "grad_norm": 1.7900562286376953, + "learning_rate": 0.0001614827062665086, + "loss": 0.1284, + "step": 37040 + }, + { + "epoch": 0.5779309914519248, + "grad_norm": 0.6629091501235962, + "learning_rate": 0.00016147230714835383, + "loss": 0.2657, + "step": 37050 + }, + { + "epoch": 0.5780869782242466, + "grad_norm": 2.502882242202759, + "learning_rate": 0.00016146190803019905, + "loss": 0.207, + "step": 37060 + }, + { + "epoch": 0.5782429649965682, + "grad_norm": 1.486069917678833, + "learning_rate": 0.00016145150891204427, + "loss": 0.1232, + "step": 37070 + }, + { + "epoch": 0.57839895176889, + "grad_norm": 0.13543163239955902, + "learning_rate": 0.0001614411097938895, + "loss": 0.3245, + "step": 37080 + }, + { + "epoch": 0.5785549385412118, + "grad_norm": 3.9703898429870605, + "learning_rate": 0.0001614307106757347, + "loss": 0.3104, + "step": 37090 + }, + { + "epoch": 0.5787109253135334, + "grad_norm": 3.2236313819885254, + "learning_rate": 0.00016142031155757993, + "loss": 0.1739, + "step": 37100 + }, + { + "epoch": 0.5788669120858552, + "grad_norm": 0.3854866325855255, + "learning_rate": 0.00016140991243942515, + "loss": 0.084, + "step": 37110 + }, + { + "epoch": 0.5790228988581768, + "grad_norm": 0.1981169581413269, + "learning_rate": 0.00016139951332127037, + "loss": 0.2367, + "step": 37120 + }, + { + "epoch": 0.5791788856304986, + "grad_norm": 2.7315785884857178, + "learning_rate": 0.0001613891142031156, + "loss": 0.1816, + "step": 37130 + }, + { + "epoch": 0.5793348724028202, + "grad_norm": 0.48251378536224365, + "learning_rate": 0.0001613787150849608, + "loss": 0.2123, + "step": 37140 + }, + { + "epoch": 0.579490859175142, + "grad_norm": 5.0431413650512695, + "learning_rate": 0.00016136831596680603, + "loss": 0.2489, + "step": 37150 + }, + { + "epoch": 0.5796468459474636, + "grad_norm": 0.81981360912323, + "learning_rate": 0.00016135791684865125, + "loss": 0.5235, + "step": 37160 + }, + { + "epoch": 0.5798028327197854, + "grad_norm": 0.5701965689659119, + "learning_rate": 0.00016134751773049647, + "loss": 0.1709, + "step": 37170 + }, + { + "epoch": 0.579958819492107, + "grad_norm": 1.3905616998672485, + "learning_rate": 0.00016133711861234169, + "loss": 0.2641, + "step": 37180 + }, + { + "epoch": 0.5801148062644288, + "grad_norm": 2.281285047531128, + "learning_rate": 0.0001613267194941869, + "loss": 0.3915, + "step": 37190 + }, + { + "epoch": 0.5802707930367504, + "grad_norm": 1.2038309574127197, + "learning_rate": 0.00016131632037603213, + "loss": 0.0881, + "step": 37200 + }, + { + "epoch": 0.5804267798090722, + "grad_norm": 0.3096400201320648, + "learning_rate": 0.00016130592125787735, + "loss": 0.1341, + "step": 37210 + }, + { + "epoch": 0.5805827665813938, + "grad_norm": 0.10219905525445938, + "learning_rate": 0.00016129552213972256, + "loss": 0.0867, + "step": 37220 + }, + { + "epoch": 0.5807387533537156, + "grad_norm": 0.13869646191596985, + "learning_rate": 0.00016128512302156778, + "loss": 0.1621, + "step": 37230 + }, + { + "epoch": 0.5808947401260373, + "grad_norm": 0.7992938160896301, + "learning_rate": 0.000161274723903413, + "loss": 0.275, + "step": 37240 + }, + { + "epoch": 0.581050726898359, + "grad_norm": 0.2635972201824188, + "learning_rate": 0.00016126432478525822, + "loss": 0.2429, + "step": 37250 + }, + { + "epoch": 0.5812067136706808, + "grad_norm": 1.7211238145828247, + "learning_rate": 0.00016125392566710344, + "loss": 0.1541, + "step": 37260 + }, + { + "epoch": 0.5813627004430024, + "grad_norm": 2.2448506355285645, + "learning_rate": 0.00016124352654894866, + "loss": 0.3689, + "step": 37270 + }, + { + "epoch": 0.5815186872153242, + "grad_norm": 1.238708257675171, + "learning_rate": 0.00016123312743079388, + "loss": 0.3292, + "step": 37280 + }, + { + "epoch": 0.5816746739876458, + "grad_norm": 0.5670503377914429, + "learning_rate": 0.0001612227283126391, + "loss": 0.1151, + "step": 37290 + }, + { + "epoch": 0.5818306607599676, + "grad_norm": 2.287304401397705, + "learning_rate": 0.00016121232919448432, + "loss": 0.3558, + "step": 37300 + }, + { + "epoch": 0.5819866475322892, + "grad_norm": 0.8009629845619202, + "learning_rate": 0.00016120193007632954, + "loss": 0.1861, + "step": 37310 + }, + { + "epoch": 0.582142634304611, + "grad_norm": 1.1901466846466064, + "learning_rate": 0.00016119153095817476, + "loss": 0.089, + "step": 37320 + }, + { + "epoch": 0.5822986210769326, + "grad_norm": 1.7198141813278198, + "learning_rate": 0.00016118113184001998, + "loss": 0.2931, + "step": 37330 + }, + { + "epoch": 0.5824546078492544, + "grad_norm": 0.8832749724388123, + "learning_rate": 0.0001611707327218652, + "loss": 0.2342, + "step": 37340 + }, + { + "epoch": 0.582610594621576, + "grad_norm": 1.0623290538787842, + "learning_rate": 0.00016116033360371042, + "loss": 0.2317, + "step": 37350 + }, + { + "epoch": 0.5827665813938978, + "grad_norm": 0.5173699855804443, + "learning_rate": 0.00016114993448555564, + "loss": 0.1299, + "step": 37360 + }, + { + "epoch": 0.5829225681662195, + "grad_norm": 0.05118720978498459, + "learning_rate": 0.00016113953536740086, + "loss": 0.0786, + "step": 37370 + }, + { + "epoch": 0.5830785549385412, + "grad_norm": 2.0060107707977295, + "learning_rate": 0.00016112913624924608, + "loss": 0.266, + "step": 37380 + }, + { + "epoch": 0.5832345417108629, + "grad_norm": 5.121693134307861, + "learning_rate": 0.0001611187371310913, + "loss": 0.1824, + "step": 37390 + }, + { + "epoch": 0.5833905284831846, + "grad_norm": 0.3568089008331299, + "learning_rate": 0.00016110833801293652, + "loss": 0.3189, + "step": 37400 + }, + { + "epoch": 0.5835465152555064, + "grad_norm": 0.8631492853164673, + "learning_rate": 0.00016109793889478174, + "loss": 0.1724, + "step": 37410 + }, + { + "epoch": 0.583702502027828, + "grad_norm": 1.9214404821395874, + "learning_rate": 0.00016108753977662693, + "loss": 0.2094, + "step": 37420 + }, + { + "epoch": 0.5838584888001498, + "grad_norm": 1.9055646657943726, + "learning_rate": 0.00016107714065847218, + "loss": 0.1258, + "step": 37430 + }, + { + "epoch": 0.5840144755724714, + "grad_norm": 0.6614194512367249, + "learning_rate": 0.00016106674154031737, + "loss": 0.1966, + "step": 37440 + }, + { + "epoch": 0.5841704623447932, + "grad_norm": 0.286883682012558, + "learning_rate": 0.00016105634242216262, + "loss": 0.588, + "step": 37450 + }, + { + "epoch": 0.5843264491171148, + "grad_norm": 0.5599405765533447, + "learning_rate": 0.0001610459433040078, + "loss": 0.3414, + "step": 37460 + }, + { + "epoch": 0.5844824358894366, + "grad_norm": 1.5908915996551514, + "learning_rate": 0.00016103554418585306, + "loss": 0.1264, + "step": 37470 + }, + { + "epoch": 0.5846384226617583, + "grad_norm": 1.0729930400848389, + "learning_rate": 0.00016102514506769825, + "loss": 0.4815, + "step": 37480 + }, + { + "epoch": 0.58479440943408, + "grad_norm": 0.8912356495857239, + "learning_rate": 0.0001610147459495435, + "loss": 0.2092, + "step": 37490 + }, + { + "epoch": 0.5849503962064017, + "grad_norm": 0.24116197228431702, + "learning_rate": 0.0001610043468313887, + "loss": 0.1552, + "step": 37500 + }, + { + "epoch": 0.5851063829787234, + "grad_norm": 2.245192527770996, + "learning_rate": 0.00016099394771323393, + "loss": 0.0553, + "step": 37510 + }, + { + "epoch": 0.5852623697510451, + "grad_norm": 0.11538795381784439, + "learning_rate": 0.00016098354859507913, + "loss": 0.0889, + "step": 37520 + }, + { + "epoch": 0.5854183565233668, + "grad_norm": 1.0372689962387085, + "learning_rate": 0.00016097314947692437, + "loss": 0.3909, + "step": 37530 + }, + { + "epoch": 0.5855743432956885, + "grad_norm": 1.1946550607681274, + "learning_rate": 0.00016096275035876957, + "loss": 0.2167, + "step": 37540 + }, + { + "epoch": 0.5857303300680102, + "grad_norm": 1.3474591970443726, + "learning_rate": 0.0001609523512406148, + "loss": 0.1561, + "step": 37550 + }, + { + "epoch": 0.585886316840332, + "grad_norm": 0.008936616592109203, + "learning_rate": 0.00016094195212246, + "loss": 0.3991, + "step": 37560 + }, + { + "epoch": 0.5860423036126536, + "grad_norm": 1.8822320699691772, + "learning_rate": 0.00016093155300430525, + "loss": 0.2438, + "step": 37570 + }, + { + "epoch": 0.5861982903849754, + "grad_norm": 1.1707524061203003, + "learning_rate": 0.00016092115388615044, + "loss": 0.308, + "step": 37580 + }, + { + "epoch": 0.586354277157297, + "grad_norm": 1.679641842842102, + "learning_rate": 0.0001609107547679957, + "loss": 0.159, + "step": 37590 + }, + { + "epoch": 0.5865102639296188, + "grad_norm": 1.6885284185409546, + "learning_rate": 0.00016090035564984088, + "loss": 0.157, + "step": 37600 + }, + { + "epoch": 0.5866662507019405, + "grad_norm": 0.3876996636390686, + "learning_rate": 0.00016088995653168613, + "loss": 0.4067, + "step": 37610 + }, + { + "epoch": 0.5868222374742622, + "grad_norm": 1.387480616569519, + "learning_rate": 0.00016087955741353132, + "loss": 0.1898, + "step": 37620 + }, + { + "epoch": 0.5869782242465839, + "grad_norm": 0.03653861582279205, + "learning_rate": 0.00016086915829537657, + "loss": 0.1652, + "step": 37630 + }, + { + "epoch": 0.5871342110189056, + "grad_norm": 0.6258412003517151, + "learning_rate": 0.00016085875917722176, + "loss": 0.2558, + "step": 37640 + }, + { + "epoch": 0.5872901977912273, + "grad_norm": 0.10519164055585861, + "learning_rate": 0.000160848360059067, + "loss": 0.243, + "step": 37650 + }, + { + "epoch": 0.587446184563549, + "grad_norm": 0.7370131015777588, + "learning_rate": 0.0001608379609409122, + "loss": 0.2343, + "step": 37660 + }, + { + "epoch": 0.5876021713358707, + "grad_norm": 0.6884517073631287, + "learning_rate": 0.00016082756182275745, + "loss": 0.1847, + "step": 37670 + }, + { + "epoch": 0.5877581581081924, + "grad_norm": 0.6326120495796204, + "learning_rate": 0.00016081716270460264, + "loss": 0.2258, + "step": 37680 + }, + { + "epoch": 0.5879141448805141, + "grad_norm": 7.697889804840088, + "learning_rate": 0.0001608067635864479, + "loss": 0.2068, + "step": 37690 + }, + { + "epoch": 0.5880701316528358, + "grad_norm": 1.8753420114517212, + "learning_rate": 0.00016079636446829308, + "loss": 0.3087, + "step": 37700 + }, + { + "epoch": 0.5882261184251576, + "grad_norm": 0.8591625094413757, + "learning_rate": 0.00016078596535013833, + "loss": 0.1082, + "step": 37710 + }, + { + "epoch": 0.5883821051974792, + "grad_norm": 0.6124423146247864, + "learning_rate": 0.00016077556623198352, + "loss": 0.1536, + "step": 37720 + }, + { + "epoch": 0.588538091969801, + "grad_norm": 0.1400165557861328, + "learning_rate": 0.00016076516711382877, + "loss": 0.145, + "step": 37730 + }, + { + "epoch": 0.5886940787421227, + "grad_norm": 1.9879004955291748, + "learning_rate": 0.00016075476799567396, + "loss": 0.4986, + "step": 37740 + }, + { + "epoch": 0.5888500655144444, + "grad_norm": 0.08687327802181244, + "learning_rate": 0.0001607443688775192, + "loss": 0.3365, + "step": 37750 + }, + { + "epoch": 0.5890060522867661, + "grad_norm": 0.3442278206348419, + "learning_rate": 0.0001607339697593644, + "loss": 0.2274, + "step": 37760 + }, + { + "epoch": 0.5891620390590878, + "grad_norm": 0.06124915927648544, + "learning_rate": 0.00016072357064120965, + "loss": 0.1915, + "step": 37770 + }, + { + "epoch": 0.5893180258314095, + "grad_norm": 1.0996816158294678, + "learning_rate": 0.00016071317152305484, + "loss": 0.1573, + "step": 37780 + }, + { + "epoch": 0.5894740126037312, + "grad_norm": 0.14738436043262482, + "learning_rate": 0.00016070277240490008, + "loss": 0.2524, + "step": 37790 + }, + { + "epoch": 0.5896299993760529, + "grad_norm": 2.3905467987060547, + "learning_rate": 0.00016069237328674528, + "loss": 0.1838, + "step": 37800 + }, + { + "epoch": 0.5897859861483746, + "grad_norm": 2.0800178050994873, + "learning_rate": 0.00016068197416859052, + "loss": 0.3749, + "step": 37810 + }, + { + "epoch": 0.5899419729206963, + "grad_norm": 3.254204511642456, + "learning_rate": 0.00016067157505043574, + "loss": 0.3777, + "step": 37820 + }, + { + "epoch": 0.590097959693018, + "grad_norm": 2.8447892665863037, + "learning_rate": 0.00016066117593228096, + "loss": 0.5338, + "step": 37830 + }, + { + "epoch": 0.5902539464653397, + "grad_norm": 0.624733030796051, + "learning_rate": 0.00016065077681412618, + "loss": 0.2331, + "step": 37840 + }, + { + "epoch": 0.5904099332376614, + "grad_norm": 0.6863507628440857, + "learning_rate": 0.0001606403776959714, + "loss": 0.1747, + "step": 37850 + }, + { + "epoch": 0.5905659200099832, + "grad_norm": 0.6785943508148193, + "learning_rate": 0.00016062997857781662, + "loss": 0.2064, + "step": 37860 + }, + { + "epoch": 0.5907219067823049, + "grad_norm": 0.5739938616752625, + "learning_rate": 0.00016061957945966181, + "loss": 0.2368, + "step": 37870 + }, + { + "epoch": 0.5908778935546266, + "grad_norm": 0.0616387277841568, + "learning_rate": 0.00016060918034150706, + "loss": 0.2224, + "step": 37880 + }, + { + "epoch": 0.5910338803269483, + "grad_norm": 2.745957136154175, + "learning_rate": 0.00016059878122335225, + "loss": 0.3161, + "step": 37890 + }, + { + "epoch": 0.59118986709927, + "grad_norm": 2.52472186088562, + "learning_rate": 0.0001605883821051975, + "loss": 0.1713, + "step": 37900 + }, + { + "epoch": 0.5913458538715917, + "grad_norm": 1.5691877603530884, + "learning_rate": 0.0001605779829870427, + "loss": 0.1402, + "step": 37910 + }, + { + "epoch": 0.5915018406439134, + "grad_norm": 1.848261833190918, + "learning_rate": 0.00016056758386888794, + "loss": 0.2305, + "step": 37920 + }, + { + "epoch": 0.5916578274162351, + "grad_norm": 0.36622482538223267, + "learning_rate": 0.00016055718475073313, + "loss": 0.1834, + "step": 37930 + }, + { + "epoch": 0.5918138141885568, + "grad_norm": 0.21201446652412415, + "learning_rate": 0.00016054678563257838, + "loss": 0.4682, + "step": 37940 + }, + { + "epoch": 0.5919698009608785, + "grad_norm": 0.38921502232551575, + "learning_rate": 0.00016053638651442357, + "loss": 0.2892, + "step": 37950 + }, + { + "epoch": 0.5921257877332002, + "grad_norm": 0.20777840912342072, + "learning_rate": 0.00016052598739626882, + "loss": 0.104, + "step": 37960 + }, + { + "epoch": 0.5922817745055219, + "grad_norm": 0.7090150713920593, + "learning_rate": 0.000160515588278114, + "loss": 0.2592, + "step": 37970 + }, + { + "epoch": 0.5924377612778436, + "grad_norm": 1.9939709901809692, + "learning_rate": 0.00016050518915995926, + "loss": 0.247, + "step": 37980 + }, + { + "epoch": 0.5925937480501653, + "grad_norm": 0.8696603178977966, + "learning_rate": 0.00016049479004180445, + "loss": 0.301, + "step": 37990 + }, + { + "epoch": 0.592749734822487, + "grad_norm": 0.4119095206260681, + "learning_rate": 0.0001604843909236497, + "loss": 0.2011, + "step": 38000 + }, + { + "epoch": 0.5929057215948088, + "grad_norm": 0.060975607484579086, + "learning_rate": 0.0001604739918054949, + "loss": 0.2286, + "step": 38010 + }, + { + "epoch": 0.5930617083671305, + "grad_norm": 0.7039875388145447, + "learning_rate": 0.00016046359268734014, + "loss": 0.0999, + "step": 38020 + }, + { + "epoch": 0.5932176951394522, + "grad_norm": 0.28402045369148254, + "learning_rate": 0.00016045319356918533, + "loss": 0.1835, + "step": 38030 + }, + { + "epoch": 0.5933736819117739, + "grad_norm": 2.217988967895508, + "learning_rate": 0.00016044279445103058, + "loss": 0.0785, + "step": 38040 + }, + { + "epoch": 0.5935296686840956, + "grad_norm": 1.4856069087982178, + "learning_rate": 0.00016043239533287577, + "loss": 0.2528, + "step": 38050 + }, + { + "epoch": 0.5936856554564173, + "grad_norm": 0.9875519871711731, + "learning_rate": 0.00016042199621472101, + "loss": 0.4445, + "step": 38060 + }, + { + "epoch": 0.593841642228739, + "grad_norm": 2.773216724395752, + "learning_rate": 0.0001604115970965662, + "loss": 0.5643, + "step": 38070 + }, + { + "epoch": 0.5939976290010607, + "grad_norm": 1.7584549188613892, + "learning_rate": 0.00016040119797841145, + "loss": 0.1446, + "step": 38080 + }, + { + "epoch": 0.5941536157733824, + "grad_norm": 1.2456908226013184, + "learning_rate": 0.00016039079886025665, + "loss": 0.1896, + "step": 38090 + }, + { + "epoch": 0.5943096025457041, + "grad_norm": 0.021098516881465912, + "learning_rate": 0.0001603803997421019, + "loss": 0.0876, + "step": 38100 + }, + { + "epoch": 0.5944655893180258, + "grad_norm": 1.8131749629974365, + "learning_rate": 0.00016037000062394709, + "loss": 0.1639, + "step": 38110 + }, + { + "epoch": 0.5946215760903475, + "grad_norm": 0.41819462180137634, + "learning_rate": 0.00016035960150579233, + "loss": 0.1311, + "step": 38120 + }, + { + "epoch": 0.5947775628626693, + "grad_norm": 0.6691219806671143, + "learning_rate": 0.00016034920238763753, + "loss": 0.1625, + "step": 38130 + }, + { + "epoch": 0.5949335496349909, + "grad_norm": 0.07854852080345154, + "learning_rate": 0.00016033880326948277, + "loss": 0.3177, + "step": 38140 + }, + { + "epoch": 0.5950895364073127, + "grad_norm": 0.7819689512252808, + "learning_rate": 0.00016032840415132796, + "loss": 0.4225, + "step": 38150 + }, + { + "epoch": 0.5952455231796344, + "grad_norm": 5.3962836265563965, + "learning_rate": 0.0001603180050331732, + "loss": 0.123, + "step": 38160 + }, + { + "epoch": 0.5954015099519561, + "grad_norm": 0.7460303902626038, + "learning_rate": 0.0001603076059150184, + "loss": 0.2182, + "step": 38170 + }, + { + "epoch": 0.5955574967242778, + "grad_norm": 0.26433685421943665, + "learning_rate": 0.00016029720679686365, + "loss": 0.1894, + "step": 38180 + }, + { + "epoch": 0.5957134834965995, + "grad_norm": 0.8701320290565491, + "learning_rate": 0.00016028680767870884, + "loss": 0.2137, + "step": 38190 + }, + { + "epoch": 0.5958694702689212, + "grad_norm": 3.9774930477142334, + "learning_rate": 0.0001602764085605541, + "loss": 0.4637, + "step": 38200 + }, + { + "epoch": 0.5960254570412429, + "grad_norm": 0.5801145434379578, + "learning_rate": 0.00016026600944239928, + "loss": 0.3036, + "step": 38210 + }, + { + "epoch": 0.5961814438135646, + "grad_norm": 1.1890723705291748, + "learning_rate": 0.00016025561032424453, + "loss": 0.0988, + "step": 38220 + }, + { + "epoch": 0.5963374305858863, + "grad_norm": 0.18521073460578918, + "learning_rate": 0.00016024521120608972, + "loss": 0.1529, + "step": 38230 + }, + { + "epoch": 0.596493417358208, + "grad_norm": 2.2602827548980713, + "learning_rate": 0.00016023481208793497, + "loss": 0.2815, + "step": 38240 + }, + { + "epoch": 0.5966494041305297, + "grad_norm": 0.4060989022254944, + "learning_rate": 0.00016022441296978016, + "loss": 0.3466, + "step": 38250 + }, + { + "epoch": 0.5968053909028515, + "grad_norm": 1.4743244647979736, + "learning_rate": 0.0001602140138516254, + "loss": 0.3561, + "step": 38260 + }, + { + "epoch": 0.5969613776751731, + "grad_norm": 5.080365180969238, + "learning_rate": 0.0001602036147334706, + "loss": 0.2914, + "step": 38270 + }, + { + "epoch": 0.5971173644474949, + "grad_norm": 0.16149812936782837, + "learning_rate": 0.00016019321561531585, + "loss": 0.1871, + "step": 38280 + }, + { + "epoch": 0.5972733512198165, + "grad_norm": 1.7535364627838135, + "learning_rate": 0.00016018281649716104, + "loss": 0.4905, + "step": 38290 + }, + { + "epoch": 0.5974293379921383, + "grad_norm": 0.4103987216949463, + "learning_rate": 0.00016017241737900629, + "loss": 0.3813, + "step": 38300 + }, + { + "epoch": 0.59758532476446, + "grad_norm": 0.5315369367599487, + "learning_rate": 0.00016016201826085148, + "loss": 0.3897, + "step": 38310 + }, + { + "epoch": 0.5977413115367817, + "grad_norm": 3.4553463459014893, + "learning_rate": 0.0001601516191426967, + "loss": 0.2047, + "step": 38320 + }, + { + "epoch": 0.5978972983091034, + "grad_norm": 2.225468397140503, + "learning_rate": 0.00016014122002454192, + "loss": 0.4028, + "step": 38330 + }, + { + "epoch": 0.5980532850814251, + "grad_norm": 2.000885486602783, + "learning_rate": 0.00016013082090638714, + "loss": 0.391, + "step": 38340 + }, + { + "epoch": 0.5982092718537468, + "grad_norm": 0.7886672616004944, + "learning_rate": 0.00016012042178823236, + "loss": 0.2994, + "step": 38350 + }, + { + "epoch": 0.5983652586260685, + "grad_norm": 0.5279917120933533, + "learning_rate": 0.00016011002267007758, + "loss": 0.1855, + "step": 38360 + }, + { + "epoch": 0.5985212453983902, + "grad_norm": 0.9789942502975464, + "learning_rate": 0.0001600996235519228, + "loss": 0.2122, + "step": 38370 + }, + { + "epoch": 0.5986772321707119, + "grad_norm": 0.6289750933647156, + "learning_rate": 0.00016008922443376802, + "loss": 0.3737, + "step": 38380 + }, + { + "epoch": 0.5988332189430337, + "grad_norm": 1.3074733018875122, + "learning_rate": 0.00016007882531561324, + "loss": 0.2382, + "step": 38390 + }, + { + "epoch": 0.5989892057153553, + "grad_norm": 1.745284914970398, + "learning_rate": 0.00016006842619745846, + "loss": 0.3167, + "step": 38400 + }, + { + "epoch": 0.5991451924876771, + "grad_norm": 0.1694452464580536, + "learning_rate": 0.00016005802707930368, + "loss": 0.1632, + "step": 38410 + }, + { + "epoch": 0.5993011792599987, + "grad_norm": 1.55560302734375, + "learning_rate": 0.0001600476279611489, + "loss": 0.2371, + "step": 38420 + }, + { + "epoch": 0.5994571660323205, + "grad_norm": 1.044997215270996, + "learning_rate": 0.00016003722884299411, + "loss": 0.0896, + "step": 38430 + }, + { + "epoch": 0.5996131528046421, + "grad_norm": 0.25250110030174255, + "learning_rate": 0.00016002682972483933, + "loss": 0.3096, + "step": 38440 + }, + { + "epoch": 0.5997691395769639, + "grad_norm": 0.1631859689950943, + "learning_rate": 0.00016001643060668455, + "loss": 0.2004, + "step": 38450 + }, + { + "epoch": 0.5999251263492856, + "grad_norm": 0.34967318177223206, + "learning_rate": 0.00016000603148852977, + "loss": 0.1476, + "step": 38460 + }, + { + "epoch": 0.6000811131216073, + "grad_norm": 1.8604000806808472, + "learning_rate": 0.000159995632370375, + "loss": 0.1884, + "step": 38470 + }, + { + "epoch": 0.600237099893929, + "grad_norm": 1.626848578453064, + "learning_rate": 0.0001599852332522202, + "loss": 0.5452, + "step": 38480 + }, + { + "epoch": 0.6003930866662507, + "grad_norm": 0.41935989260673523, + "learning_rate": 0.00015997483413406543, + "loss": 0.3047, + "step": 38490 + }, + { + "epoch": 0.6005490734385724, + "grad_norm": 1.355086088180542, + "learning_rate": 0.00015996443501591065, + "loss": 0.1246, + "step": 38500 + }, + { + "epoch": 0.6007050602108941, + "grad_norm": 1.5382829904556274, + "learning_rate": 0.00015995403589775587, + "loss": 0.2324, + "step": 38510 + }, + { + "epoch": 0.6008610469832159, + "grad_norm": 1.0250797271728516, + "learning_rate": 0.0001599436367796011, + "loss": 0.181, + "step": 38520 + }, + { + "epoch": 0.6010170337555375, + "grad_norm": 0.07823364436626434, + "learning_rate": 0.0001599332376614463, + "loss": 0.1492, + "step": 38530 + }, + { + "epoch": 0.6011730205278593, + "grad_norm": 0.07573071867227554, + "learning_rate": 0.00015992283854329153, + "loss": 0.1256, + "step": 38540 + }, + { + "epoch": 0.6013290073001809, + "grad_norm": 0.5608569383621216, + "learning_rate": 0.00015991243942513675, + "loss": 0.378, + "step": 38550 + }, + { + "epoch": 0.6014849940725027, + "grad_norm": 0.7001075744628906, + "learning_rate": 0.00015990204030698197, + "loss": 0.2518, + "step": 38560 + }, + { + "epoch": 0.6016409808448243, + "grad_norm": 4.868145942687988, + "learning_rate": 0.0001598916411888272, + "loss": 0.1857, + "step": 38570 + }, + { + "epoch": 0.6017969676171461, + "grad_norm": 0.05653705075383186, + "learning_rate": 0.0001598812420706724, + "loss": 0.1918, + "step": 38580 + }, + { + "epoch": 0.6019529543894677, + "grad_norm": 0.8556832671165466, + "learning_rate": 0.00015987084295251763, + "loss": 0.1926, + "step": 38590 + }, + { + "epoch": 0.6021089411617895, + "grad_norm": 1.1988335847854614, + "learning_rate": 0.00015986044383436285, + "loss": 0.3884, + "step": 38600 + }, + { + "epoch": 0.6022649279341112, + "grad_norm": 1.5978832244873047, + "learning_rate": 0.00015985004471620807, + "loss": 0.1986, + "step": 38610 + }, + { + "epoch": 0.6024209147064329, + "grad_norm": 1.084174633026123, + "learning_rate": 0.0001598396455980533, + "loss": 0.4602, + "step": 38620 + }, + { + "epoch": 0.6025769014787546, + "grad_norm": 1.5116629600524902, + "learning_rate": 0.0001598292464798985, + "loss": 0.292, + "step": 38630 + }, + { + "epoch": 0.6027328882510763, + "grad_norm": 0.23945322632789612, + "learning_rate": 0.00015981884736174373, + "loss": 0.1904, + "step": 38640 + }, + { + "epoch": 0.602888875023398, + "grad_norm": 0.348156601190567, + "learning_rate": 0.00015980844824358895, + "loss": 0.1442, + "step": 38650 + }, + { + "epoch": 0.6030448617957197, + "grad_norm": 1.876736044883728, + "learning_rate": 0.00015979804912543417, + "loss": 0.248, + "step": 38660 + }, + { + "epoch": 0.6032008485680415, + "grad_norm": 0.2564373016357422, + "learning_rate": 0.00015978765000727939, + "loss": 0.2429, + "step": 38670 + }, + { + "epoch": 0.6033568353403631, + "grad_norm": 1.824837327003479, + "learning_rate": 0.0001597772508891246, + "loss": 0.2593, + "step": 38680 + }, + { + "epoch": 0.6035128221126849, + "grad_norm": 1.6967028379440308, + "learning_rate": 0.00015976685177096985, + "loss": 0.3918, + "step": 38690 + }, + { + "epoch": 0.6036688088850065, + "grad_norm": 1.6629718542099, + "learning_rate": 0.00015975645265281504, + "loss": 0.2529, + "step": 38700 + }, + { + "epoch": 0.6038247956573283, + "grad_norm": 0.19047370553016663, + "learning_rate": 0.0001597460535346603, + "loss": 0.4545, + "step": 38710 + }, + { + "epoch": 0.6039807824296499, + "grad_norm": 0.6910511255264282, + "learning_rate": 0.00015973565441650548, + "loss": 0.2226, + "step": 38720 + }, + { + "epoch": 0.6041367692019717, + "grad_norm": 1.9387328624725342, + "learning_rate": 0.00015972525529835073, + "loss": 0.5561, + "step": 38730 + }, + { + "epoch": 0.6042927559742933, + "grad_norm": 1.4491708278656006, + "learning_rate": 0.00015971485618019592, + "loss": 0.2987, + "step": 38740 + }, + { + "epoch": 0.6044487427466151, + "grad_norm": 0.32534459233283997, + "learning_rate": 0.00015970445706204117, + "loss": 0.1683, + "step": 38750 + }, + { + "epoch": 0.6046047295189368, + "grad_norm": 0.17335011065006256, + "learning_rate": 0.00015969405794388636, + "loss": 0.1554, + "step": 38760 + }, + { + "epoch": 0.6047607162912585, + "grad_norm": 0.39481663703918457, + "learning_rate": 0.00015968365882573158, + "loss": 0.186, + "step": 38770 + }, + { + "epoch": 0.6049167030635803, + "grad_norm": 0.10422719269990921, + "learning_rate": 0.0001596732597075768, + "loss": 0.1599, + "step": 38780 + }, + { + "epoch": 0.6050726898359019, + "grad_norm": 2.288064479827881, + "learning_rate": 0.00015966286058942202, + "loss": 0.1944, + "step": 38790 + }, + { + "epoch": 0.6052286766082237, + "grad_norm": 1.0062401294708252, + "learning_rate": 0.00015965246147126724, + "loss": 0.176, + "step": 38800 + }, + { + "epoch": 0.6053846633805453, + "grad_norm": 0.21470843255519867, + "learning_rate": 0.00015964206235311246, + "loss": 0.183, + "step": 38810 + }, + { + "epoch": 0.6055406501528671, + "grad_norm": 2.1730270385742188, + "learning_rate": 0.00015963166323495768, + "loss": 0.1788, + "step": 38820 + }, + { + "epoch": 0.6056966369251887, + "grad_norm": 0.6188749074935913, + "learning_rate": 0.0001596212641168029, + "loss": 0.2746, + "step": 38830 + }, + { + "epoch": 0.6058526236975105, + "grad_norm": 1.283058524131775, + "learning_rate": 0.00015961086499864812, + "loss": 0.1066, + "step": 38840 + }, + { + "epoch": 0.6060086104698321, + "grad_norm": 1.1469789743423462, + "learning_rate": 0.00015960046588049334, + "loss": 0.0692, + "step": 38850 + }, + { + "epoch": 0.6061645972421539, + "grad_norm": 1.9450418949127197, + "learning_rate": 0.00015959006676233856, + "loss": 0.2154, + "step": 38860 + }, + { + "epoch": 0.6063205840144755, + "grad_norm": 7.193401336669922, + "learning_rate": 0.00015957966764418378, + "loss": 0.218, + "step": 38870 + }, + { + "epoch": 0.6064765707867973, + "grad_norm": 3.5315003395080566, + "learning_rate": 0.000159569268526029, + "loss": 0.312, + "step": 38880 + }, + { + "epoch": 0.6066325575591189, + "grad_norm": 1.2873650789260864, + "learning_rate": 0.00015955886940787422, + "loss": 0.2587, + "step": 38890 + }, + { + "epoch": 0.6067885443314407, + "grad_norm": 1.1193941831588745, + "learning_rate": 0.00015954847028971944, + "loss": 0.2358, + "step": 38900 + }, + { + "epoch": 0.6069445311037625, + "grad_norm": 1.0471165180206299, + "learning_rate": 0.00015953807117156466, + "loss": 0.2639, + "step": 38910 + }, + { + "epoch": 0.6071005178760841, + "grad_norm": 1.7760034799575806, + "learning_rate": 0.00015952767205340988, + "loss": 0.248, + "step": 38920 + }, + { + "epoch": 0.6072565046484059, + "grad_norm": 0.050742994993925095, + "learning_rate": 0.0001595172729352551, + "loss": 0.1287, + "step": 38930 + }, + { + "epoch": 0.6074124914207275, + "grad_norm": 0.03016967698931694, + "learning_rate": 0.00015950687381710032, + "loss": 0.1843, + "step": 38940 + }, + { + "epoch": 0.6075684781930493, + "grad_norm": 0.6029847860336304, + "learning_rate": 0.00015949647469894554, + "loss": 0.2294, + "step": 38950 + }, + { + "epoch": 0.6077244649653709, + "grad_norm": 0.2955610454082489, + "learning_rate": 0.00015948607558079076, + "loss": 0.3336, + "step": 38960 + }, + { + "epoch": 0.6078804517376927, + "grad_norm": 1.9104161262512207, + "learning_rate": 0.00015947567646263597, + "loss": 0.189, + "step": 38970 + }, + { + "epoch": 0.6080364385100143, + "grad_norm": 0.09197711199522018, + "learning_rate": 0.0001594652773444812, + "loss": 0.1612, + "step": 38980 + }, + { + "epoch": 0.6081924252823361, + "grad_norm": 1.4233628511428833, + "learning_rate": 0.00015945487822632641, + "loss": 0.1521, + "step": 38990 + }, + { + "epoch": 0.6083484120546577, + "grad_norm": 9.787493705749512, + "learning_rate": 0.00015944447910817163, + "loss": 0.135, + "step": 39000 + }, + { + "epoch": 0.6085043988269795, + "grad_norm": 1.3848742246627808, + "learning_rate": 0.00015943407999001685, + "loss": 0.2268, + "step": 39010 + }, + { + "epoch": 0.6086603855993011, + "grad_norm": 0.1524626612663269, + "learning_rate": 0.00015942368087186207, + "loss": 0.3154, + "step": 39020 + }, + { + "epoch": 0.6088163723716229, + "grad_norm": 2.1885435581207275, + "learning_rate": 0.0001594132817537073, + "loss": 0.3174, + "step": 39030 + }, + { + "epoch": 0.6089723591439445, + "grad_norm": 1.4719898700714111, + "learning_rate": 0.0001594028826355525, + "loss": 0.2959, + "step": 39040 + }, + { + "epoch": 0.6091283459162663, + "grad_norm": 0.7844040989875793, + "learning_rate": 0.00015939248351739773, + "loss": 0.3021, + "step": 39050 + }, + { + "epoch": 0.6092843326885881, + "grad_norm": 0.17660890519618988, + "learning_rate": 0.00015938208439924295, + "loss": 0.3608, + "step": 39060 + }, + { + "epoch": 0.6094403194609097, + "grad_norm": 0.24621741473674774, + "learning_rate": 0.00015937168528108817, + "loss": 0.1461, + "step": 39070 + }, + { + "epoch": 0.6095963062332315, + "grad_norm": 0.7903050184249878, + "learning_rate": 0.0001593612861629334, + "loss": 0.1195, + "step": 39080 + }, + { + "epoch": 0.6097522930055531, + "grad_norm": 0.04043950140476227, + "learning_rate": 0.0001593508870447786, + "loss": 0.1938, + "step": 39090 + }, + { + "epoch": 0.6099082797778749, + "grad_norm": 0.7602111101150513, + "learning_rate": 0.00015934048792662383, + "loss": 0.1732, + "step": 39100 + }, + { + "epoch": 0.6100642665501965, + "grad_norm": 0.7902756333351135, + "learning_rate": 0.00015933008880846905, + "loss": 0.2351, + "step": 39110 + }, + { + "epoch": 0.6102202533225183, + "grad_norm": 1.7035608291625977, + "learning_rate": 0.00015931968969031427, + "loss": 0.2409, + "step": 39120 + }, + { + "epoch": 0.6103762400948399, + "grad_norm": 0.7998217940330505, + "learning_rate": 0.0001593092905721595, + "loss": 0.3862, + "step": 39130 + }, + { + "epoch": 0.6105322268671617, + "grad_norm": 0.3530103862285614, + "learning_rate": 0.0001592988914540047, + "loss": 0.2176, + "step": 39140 + }, + { + "epoch": 0.6106882136394833, + "grad_norm": 0.9218409657478333, + "learning_rate": 0.00015928849233584993, + "loss": 0.1362, + "step": 39150 + }, + { + "epoch": 0.6108442004118051, + "grad_norm": 0.643829345703125, + "learning_rate": 0.00015927809321769515, + "loss": 0.2309, + "step": 39160 + }, + { + "epoch": 0.6110001871841267, + "grad_norm": 0.1810176521539688, + "learning_rate": 0.00015926769409954037, + "loss": 0.2461, + "step": 39170 + }, + { + "epoch": 0.6111561739564485, + "grad_norm": 1.3435649871826172, + "learning_rate": 0.0001592572949813856, + "loss": 0.2937, + "step": 39180 + }, + { + "epoch": 0.6113121607287701, + "grad_norm": 1.1048580408096313, + "learning_rate": 0.0001592468958632308, + "loss": 0.2983, + "step": 39190 + }, + { + "epoch": 0.6114681475010919, + "grad_norm": 0.09463644027709961, + "learning_rate": 0.00015923649674507603, + "loss": 0.2318, + "step": 39200 + }, + { + "epoch": 0.6116241342734137, + "grad_norm": 0.9329594969749451, + "learning_rate": 0.00015922609762692125, + "loss": 0.3095, + "step": 39210 + }, + { + "epoch": 0.6117801210457353, + "grad_norm": 1.8576663732528687, + "learning_rate": 0.00015921569850876647, + "loss": 0.2336, + "step": 39220 + }, + { + "epoch": 0.6119361078180571, + "grad_norm": 1.6159908771514893, + "learning_rate": 0.00015920529939061169, + "loss": 0.1745, + "step": 39230 + }, + { + "epoch": 0.6120920945903787, + "grad_norm": 1.289712905883789, + "learning_rate": 0.0001591949002724569, + "loss": 0.2031, + "step": 39240 + }, + { + "epoch": 0.6122480813627005, + "grad_norm": 0.4183761179447174, + "learning_rate": 0.00015918450115430212, + "loss": 0.2911, + "step": 39250 + }, + { + "epoch": 0.6124040681350221, + "grad_norm": 2.125042200088501, + "learning_rate": 0.00015917410203614734, + "loss": 0.1086, + "step": 39260 + }, + { + "epoch": 0.6125600549073439, + "grad_norm": 3.527683973312378, + "learning_rate": 0.00015916370291799256, + "loss": 0.3356, + "step": 39270 + }, + { + "epoch": 0.6127160416796655, + "grad_norm": 1.1525741815567017, + "learning_rate": 0.00015915330379983778, + "loss": 0.222, + "step": 39280 + }, + { + "epoch": 0.6128720284519873, + "grad_norm": 0.5210650563240051, + "learning_rate": 0.000159142904681683, + "loss": 0.1563, + "step": 39290 + }, + { + "epoch": 0.6130280152243089, + "grad_norm": 0.2881133258342743, + "learning_rate": 0.00015913250556352822, + "loss": 0.346, + "step": 39300 + }, + { + "epoch": 0.6131840019966307, + "grad_norm": 1.7034498453140259, + "learning_rate": 0.00015912210644537344, + "loss": 0.1822, + "step": 39310 + }, + { + "epoch": 0.6133399887689523, + "grad_norm": 2.6305196285247803, + "learning_rate": 0.00015911170732721866, + "loss": 0.3373, + "step": 39320 + }, + { + "epoch": 0.6134959755412741, + "grad_norm": 0.4541589915752411, + "learning_rate": 0.00015910130820906388, + "loss": 0.2146, + "step": 39330 + }, + { + "epoch": 0.6136519623135958, + "grad_norm": 0.03301222622394562, + "learning_rate": 0.0001590909090909091, + "loss": 0.1787, + "step": 39340 + }, + { + "epoch": 0.6138079490859175, + "grad_norm": 0.7106437087059021, + "learning_rate": 0.00015908050997275432, + "loss": 0.2315, + "step": 39350 + }, + { + "epoch": 0.6139639358582393, + "grad_norm": 0.381320059299469, + "learning_rate": 0.00015907011085459954, + "loss": 0.1917, + "step": 39360 + }, + { + "epoch": 0.6141199226305609, + "grad_norm": 0.5418739914894104, + "learning_rate": 0.00015905971173644476, + "loss": 0.1571, + "step": 39370 + }, + { + "epoch": 0.6142759094028827, + "grad_norm": 0.3135250210762024, + "learning_rate": 0.00015904931261828998, + "loss": 0.0858, + "step": 39380 + }, + { + "epoch": 0.6144318961752043, + "grad_norm": 1.6296218633651733, + "learning_rate": 0.0001590389135001352, + "loss": 0.2297, + "step": 39390 + }, + { + "epoch": 0.6145878829475261, + "grad_norm": 0.5768391489982605, + "learning_rate": 0.00015902851438198042, + "loss": 0.1699, + "step": 39400 + }, + { + "epoch": 0.6147438697198477, + "grad_norm": 25.794370651245117, + "learning_rate": 0.00015901811526382564, + "loss": 0.1903, + "step": 39410 + }, + { + "epoch": 0.6148998564921695, + "grad_norm": 0.46698322892189026, + "learning_rate": 0.00015900771614567086, + "loss": 0.1415, + "step": 39420 + }, + { + "epoch": 0.6150558432644911, + "grad_norm": 1.3042336702346802, + "learning_rate": 0.00015899731702751608, + "loss": 0.1604, + "step": 39430 + }, + { + "epoch": 0.6152118300368129, + "grad_norm": 1.2311556339263916, + "learning_rate": 0.0001589869179093613, + "loss": 0.2555, + "step": 39440 + }, + { + "epoch": 0.6153678168091345, + "grad_norm": 0.6290935277938843, + "learning_rate": 0.00015897651879120652, + "loss": 0.1723, + "step": 39450 + }, + { + "epoch": 0.6155238035814563, + "grad_norm": 1.4786419868469238, + "learning_rate": 0.00015896611967305174, + "loss": 0.3717, + "step": 39460 + }, + { + "epoch": 0.615679790353778, + "grad_norm": 0.7589457631111145, + "learning_rate": 0.00015895572055489696, + "loss": 0.1912, + "step": 39470 + }, + { + "epoch": 0.6158357771260997, + "grad_norm": 0.10508652776479721, + "learning_rate": 0.00015894532143674218, + "loss": 0.1353, + "step": 39480 + }, + { + "epoch": 0.6159917638984214, + "grad_norm": 0.07394446432590485, + "learning_rate": 0.0001589349223185874, + "loss": 0.1195, + "step": 39490 + }, + { + "epoch": 0.6161477506707431, + "grad_norm": 0.11420007050037384, + "learning_rate": 0.00015892452320043262, + "loss": 0.0976, + "step": 39500 + }, + { + "epoch": 0.6163037374430649, + "grad_norm": 1.2442359924316406, + "learning_rate": 0.00015891412408227784, + "loss": 0.1084, + "step": 39510 + }, + { + "epoch": 0.6164597242153865, + "grad_norm": 0.4080588221549988, + "learning_rate": 0.00015890372496412306, + "loss": 0.2191, + "step": 39520 + }, + { + "epoch": 0.6166157109877083, + "grad_norm": 0.25214430689811707, + "learning_rate": 0.00015889332584596827, + "loss": 0.5548, + "step": 39530 + }, + { + "epoch": 0.6167716977600299, + "grad_norm": 0.41108590364456177, + "learning_rate": 0.0001588829267278135, + "loss": 0.4213, + "step": 39540 + }, + { + "epoch": 0.6169276845323517, + "grad_norm": 3.3010945320129395, + "learning_rate": 0.00015887252760965871, + "loss": 0.3946, + "step": 39550 + }, + { + "epoch": 0.6170836713046733, + "grad_norm": 0.25975245237350464, + "learning_rate": 0.00015886212849150393, + "loss": 0.3045, + "step": 39560 + }, + { + "epoch": 0.6172396580769951, + "grad_norm": 1.255585789680481, + "learning_rate": 0.00015885172937334915, + "loss": 0.2249, + "step": 39570 + }, + { + "epoch": 0.6173956448493167, + "grad_norm": 1.9135524034500122, + "learning_rate": 0.00015884133025519437, + "loss": 0.2401, + "step": 39580 + }, + { + "epoch": 0.6175516316216385, + "grad_norm": 2.1866586208343506, + "learning_rate": 0.0001588309311370396, + "loss": 0.1726, + "step": 39590 + }, + { + "epoch": 0.6177076183939602, + "grad_norm": 2.487704277038574, + "learning_rate": 0.0001588205320188848, + "loss": 0.2728, + "step": 39600 + }, + { + "epoch": 0.6178636051662819, + "grad_norm": 0.7800028920173645, + "learning_rate": 0.00015881013290073003, + "loss": 0.1522, + "step": 39610 + }, + { + "epoch": 0.6180195919386036, + "grad_norm": 3.468012571334839, + "learning_rate": 0.00015879973378257525, + "loss": 0.1984, + "step": 39620 + }, + { + "epoch": 0.6181755787109253, + "grad_norm": 1.5809861421585083, + "learning_rate": 0.00015878933466442047, + "loss": 0.2632, + "step": 39630 + }, + { + "epoch": 0.618331565483247, + "grad_norm": 0.5371276140213013, + "learning_rate": 0.0001587789355462657, + "loss": 0.1915, + "step": 39640 + }, + { + "epoch": 0.6184875522555687, + "grad_norm": 4.711477279663086, + "learning_rate": 0.0001587685364281109, + "loss": 0.5271, + "step": 39650 + }, + { + "epoch": 0.6186435390278905, + "grad_norm": 1.5581507682800293, + "learning_rate": 0.00015875813730995613, + "loss": 0.2482, + "step": 39660 + }, + { + "epoch": 0.6187995258002121, + "grad_norm": 1.7416914701461792, + "learning_rate": 0.00015874773819180132, + "loss": 0.2018, + "step": 39670 + }, + { + "epoch": 0.6189555125725339, + "grad_norm": 0.6761413216590881, + "learning_rate": 0.00015873733907364657, + "loss": 0.1048, + "step": 39680 + }, + { + "epoch": 0.6191114993448555, + "grad_norm": 1.2344286441802979, + "learning_rate": 0.00015872693995549176, + "loss": 0.1492, + "step": 39690 + }, + { + "epoch": 0.6192674861171773, + "grad_norm": 4.924708843231201, + "learning_rate": 0.000158716540837337, + "loss": 0.2234, + "step": 39700 + }, + { + "epoch": 0.619423472889499, + "grad_norm": 2.3538808822631836, + "learning_rate": 0.0001587061417191822, + "loss": 0.3405, + "step": 39710 + }, + { + "epoch": 0.6195794596618207, + "grad_norm": 0.24888922274112701, + "learning_rate": 0.00015869574260102745, + "loss": 0.1021, + "step": 39720 + }, + { + "epoch": 0.6197354464341424, + "grad_norm": 1.2719457149505615, + "learning_rate": 0.00015868534348287264, + "loss": 0.2832, + "step": 39730 + }, + { + "epoch": 0.6198914332064641, + "grad_norm": 2.80794358253479, + "learning_rate": 0.0001586749443647179, + "loss": 0.2295, + "step": 39740 + }, + { + "epoch": 0.6200474199787858, + "grad_norm": 0.09828778356313705, + "learning_rate": 0.00015866454524656308, + "loss": 0.3574, + "step": 39750 + }, + { + "epoch": 0.6202034067511075, + "grad_norm": 1.0614128112792969, + "learning_rate": 0.00015865414612840833, + "loss": 0.1886, + "step": 39760 + }, + { + "epoch": 0.6203593935234292, + "grad_norm": 0.6939805150032043, + "learning_rate": 0.00015864374701025352, + "loss": 0.3412, + "step": 39770 + }, + { + "epoch": 0.6205153802957509, + "grad_norm": 0.06902116537094116, + "learning_rate": 0.00015863334789209877, + "loss": 0.1647, + "step": 39780 + }, + { + "epoch": 0.6206713670680726, + "grad_norm": 2.05729079246521, + "learning_rate": 0.00015862294877394396, + "loss": 0.168, + "step": 39790 + }, + { + "epoch": 0.6208273538403943, + "grad_norm": 4.69445276260376, + "learning_rate": 0.0001586125496557892, + "loss": 0.1701, + "step": 39800 + }, + { + "epoch": 0.6209833406127161, + "grad_norm": 0.6905967593193054, + "learning_rate": 0.0001586021505376344, + "loss": 0.2823, + "step": 39810 + }, + { + "epoch": 0.6211393273850377, + "grad_norm": 3.0361857414245605, + "learning_rate": 0.00015859175141947964, + "loss": 0.1839, + "step": 39820 + }, + { + "epoch": 0.6212953141573595, + "grad_norm": 0.08303021639585495, + "learning_rate": 0.00015858135230132484, + "loss": 0.2183, + "step": 39830 + }, + { + "epoch": 0.6214513009296811, + "grad_norm": 3.230077028274536, + "learning_rate": 0.00015857095318317008, + "loss": 0.1727, + "step": 39840 + }, + { + "epoch": 0.6216072877020029, + "grad_norm": 1.2884387969970703, + "learning_rate": 0.00015856055406501528, + "loss": 0.1081, + "step": 39850 + }, + { + "epoch": 0.6217632744743246, + "grad_norm": 2.912475109100342, + "learning_rate": 0.00015855015494686052, + "loss": 0.2695, + "step": 39860 + }, + { + "epoch": 0.6219192612466463, + "grad_norm": 0.14122223854064941, + "learning_rate": 0.00015853975582870572, + "loss": 0.3006, + "step": 39870 + }, + { + "epoch": 0.622075248018968, + "grad_norm": 0.8160560131072998, + "learning_rate": 0.00015852935671055096, + "loss": 0.1204, + "step": 39880 + }, + { + "epoch": 0.6222312347912897, + "grad_norm": 1.6930943727493286, + "learning_rate": 0.00015851895759239615, + "loss": 0.1809, + "step": 39890 + }, + { + "epoch": 0.6223872215636114, + "grad_norm": 0.008987984620034695, + "learning_rate": 0.0001585085584742414, + "loss": 0.125, + "step": 39900 + }, + { + "epoch": 0.6225432083359331, + "grad_norm": 2.6068177223205566, + "learning_rate": 0.0001584981593560866, + "loss": 0.3131, + "step": 39910 + }, + { + "epoch": 0.6226991951082548, + "grad_norm": 0.5021041035652161, + "learning_rate": 0.00015848776023793184, + "loss": 0.2023, + "step": 39920 + }, + { + "epoch": 0.6228551818805765, + "grad_norm": 1.8558531999588013, + "learning_rate": 0.00015847736111977703, + "loss": 0.3135, + "step": 39930 + }, + { + "epoch": 0.6230111686528982, + "grad_norm": 0.019155239686369896, + "learning_rate": 0.00015846696200162228, + "loss": 0.2282, + "step": 39940 + }, + { + "epoch": 0.6231671554252199, + "grad_norm": 0.47993865609169006, + "learning_rate": 0.00015845656288346747, + "loss": 0.1143, + "step": 39950 + }, + { + "epoch": 0.6233231421975417, + "grad_norm": 2.553943157196045, + "learning_rate": 0.00015844616376531272, + "loss": 0.2061, + "step": 39960 + }, + { + "epoch": 0.6234791289698633, + "grad_norm": 0.22600245475769043, + "learning_rate": 0.0001584357646471579, + "loss": 0.1808, + "step": 39970 + }, + { + "epoch": 0.6236351157421851, + "grad_norm": 0.7153030037879944, + "learning_rate": 0.00015842536552900316, + "loss": 0.2221, + "step": 39980 + }, + { + "epoch": 0.6237911025145068, + "grad_norm": 0.12053301930427551, + "learning_rate": 0.00015841496641084835, + "loss": 0.2349, + "step": 39990 + }, + { + "epoch": 0.6239470892868285, + "grad_norm": 0.44607964158058167, + "learning_rate": 0.0001584045672926936, + "loss": 0.234, + "step": 40000 + }, + { + "epoch": 0.6241030760591502, + "grad_norm": 4.0730366706848145, + "learning_rate": 0.0001583941681745388, + "loss": 0.3213, + "step": 40010 + }, + { + "epoch": 0.6242590628314719, + "grad_norm": 2.8753201961517334, + "learning_rate": 0.00015838376905638404, + "loss": 0.1745, + "step": 40020 + }, + { + "epoch": 0.6244150496037936, + "grad_norm": 1.7546380758285522, + "learning_rate": 0.00015837336993822926, + "loss": 0.2564, + "step": 40030 + }, + { + "epoch": 0.6245710363761153, + "grad_norm": 2.7480337619781494, + "learning_rate": 0.00015836297082007448, + "loss": 0.2722, + "step": 40040 + }, + { + "epoch": 0.624727023148437, + "grad_norm": 1.9213544130325317, + "learning_rate": 0.0001583525717019197, + "loss": 0.3495, + "step": 40050 + }, + { + "epoch": 0.6248830099207587, + "grad_norm": 1.5569000244140625, + "learning_rate": 0.00015834217258376492, + "loss": 0.1744, + "step": 40060 + }, + { + "epoch": 0.6250389966930804, + "grad_norm": 0.4767601788043976, + "learning_rate": 0.00015833177346561014, + "loss": 0.1739, + "step": 40070 + }, + { + "epoch": 0.6251949834654021, + "grad_norm": 3.514345407485962, + "learning_rate": 0.00015832137434745536, + "loss": 0.3174, + "step": 40080 + }, + { + "epoch": 0.6253509702377238, + "grad_norm": 0.7669044137001038, + "learning_rate": 0.00015831097522930057, + "loss": 0.1893, + "step": 40090 + }, + { + "epoch": 0.6255069570100456, + "grad_norm": 0.3902212083339691, + "learning_rate": 0.0001583005761111458, + "loss": 0.2533, + "step": 40100 + }, + { + "epoch": 0.6256629437823673, + "grad_norm": 1.3150569200515747, + "learning_rate": 0.00015829017699299101, + "loss": 0.2491, + "step": 40110 + }, + { + "epoch": 0.625818930554689, + "grad_norm": 0.9833875894546509, + "learning_rate": 0.00015827977787483623, + "loss": 0.2889, + "step": 40120 + }, + { + "epoch": 0.6259749173270107, + "grad_norm": 1.0952039957046509, + "learning_rate": 0.00015826937875668145, + "loss": 0.2213, + "step": 40130 + }, + { + "epoch": 0.6261309040993324, + "grad_norm": 0.32410308718681335, + "learning_rate": 0.00015825897963852665, + "loss": 0.1123, + "step": 40140 + }, + { + "epoch": 0.6262868908716541, + "grad_norm": 0.05873742699623108, + "learning_rate": 0.0001582485805203719, + "loss": 0.1944, + "step": 40150 + }, + { + "epoch": 0.6264428776439758, + "grad_norm": 0.936841607093811, + "learning_rate": 0.00015823818140221709, + "loss": 0.2605, + "step": 40160 + }, + { + "epoch": 0.6265988644162975, + "grad_norm": 3.288151264190674, + "learning_rate": 0.00015822778228406233, + "loss": 0.201, + "step": 40170 + }, + { + "epoch": 0.6267548511886192, + "grad_norm": 0.45985645055770874, + "learning_rate": 0.00015821738316590752, + "loss": 0.094, + "step": 40180 + }, + { + "epoch": 0.6269108379609409, + "grad_norm": 3.8212318420410156, + "learning_rate": 0.00015820698404775277, + "loss": 0.2554, + "step": 40190 + }, + { + "epoch": 0.6270668247332626, + "grad_norm": 1.8019777536392212, + "learning_rate": 0.00015819658492959796, + "loss": 0.5479, + "step": 40200 + }, + { + "epoch": 0.6272228115055843, + "grad_norm": 0.49328944087028503, + "learning_rate": 0.0001581861858114432, + "loss": 0.2608, + "step": 40210 + }, + { + "epoch": 0.627378798277906, + "grad_norm": 0.30712059140205383, + "learning_rate": 0.0001581757866932884, + "loss": 0.235, + "step": 40220 + }, + { + "epoch": 0.6275347850502278, + "grad_norm": 3.460827112197876, + "learning_rate": 0.00015816538757513365, + "loss": 0.3527, + "step": 40230 + }, + { + "epoch": 0.6276907718225494, + "grad_norm": 1.6247568130493164, + "learning_rate": 0.00015815498845697884, + "loss": 0.3031, + "step": 40240 + }, + { + "epoch": 0.6278467585948712, + "grad_norm": 0.5068777203559875, + "learning_rate": 0.0001581445893388241, + "loss": 0.3529, + "step": 40250 + }, + { + "epoch": 0.6280027453671929, + "grad_norm": 0.2704099416732788, + "learning_rate": 0.00015813419022066928, + "loss": 0.1043, + "step": 40260 + }, + { + "epoch": 0.6281587321395146, + "grad_norm": 0.07768986374139786, + "learning_rate": 0.00015812379110251453, + "loss": 0.257, + "step": 40270 + }, + { + "epoch": 0.6283147189118363, + "grad_norm": 2.6940250396728516, + "learning_rate": 0.00015811339198435972, + "loss": 0.1694, + "step": 40280 + }, + { + "epoch": 0.628470705684158, + "grad_norm": 2.3470752239227295, + "learning_rate": 0.00015810299286620497, + "loss": 0.2196, + "step": 40290 + }, + { + "epoch": 0.6286266924564797, + "grad_norm": 1.6333197355270386, + "learning_rate": 0.00015809259374805016, + "loss": 0.2452, + "step": 40300 + }, + { + "epoch": 0.6287826792288014, + "grad_norm": 0.28753232955932617, + "learning_rate": 0.0001580821946298954, + "loss": 0.0601, + "step": 40310 + }, + { + "epoch": 0.6289386660011231, + "grad_norm": 0.8923532366752625, + "learning_rate": 0.0001580717955117406, + "loss": 0.455, + "step": 40320 + }, + { + "epoch": 0.6290946527734448, + "grad_norm": 2.404088020324707, + "learning_rate": 0.00015806139639358585, + "loss": 0.466, + "step": 40330 + }, + { + "epoch": 0.6292506395457665, + "grad_norm": 1.0499215126037598, + "learning_rate": 0.00015805099727543104, + "loss": 0.4372, + "step": 40340 + }, + { + "epoch": 0.6294066263180882, + "grad_norm": 0.21907389163970947, + "learning_rate": 0.00015804059815727629, + "loss": 0.1628, + "step": 40350 + }, + { + "epoch": 0.62956261309041, + "grad_norm": 1.1540305614471436, + "learning_rate": 0.00015803019903912148, + "loss": 0.1888, + "step": 40360 + }, + { + "epoch": 0.6297185998627316, + "grad_norm": 0.44641104340553284, + "learning_rate": 0.00015801979992096672, + "loss": 0.2372, + "step": 40370 + }, + { + "epoch": 0.6298745866350534, + "grad_norm": 1.5536185503005981, + "learning_rate": 0.00015800940080281192, + "loss": 0.0855, + "step": 40380 + }, + { + "epoch": 0.630030573407375, + "grad_norm": 5.066720485687256, + "learning_rate": 0.00015799900168465716, + "loss": 0.3324, + "step": 40390 + }, + { + "epoch": 0.6301865601796968, + "grad_norm": 1.0257264375686646, + "learning_rate": 0.00015798860256650236, + "loss": 0.1872, + "step": 40400 + }, + { + "epoch": 0.6303425469520185, + "grad_norm": 1.108614206314087, + "learning_rate": 0.0001579782034483476, + "loss": 0.3244, + "step": 40410 + }, + { + "epoch": 0.6304985337243402, + "grad_norm": 0.9430763125419617, + "learning_rate": 0.0001579678043301928, + "loss": 0.2388, + "step": 40420 + }, + { + "epoch": 0.6306545204966619, + "grad_norm": 0.3578563332557678, + "learning_rate": 0.00015795740521203804, + "loss": 0.3072, + "step": 40430 + }, + { + "epoch": 0.6308105072689836, + "grad_norm": 0.6251798272132874, + "learning_rate": 0.00015794700609388323, + "loss": 0.3246, + "step": 40440 + }, + { + "epoch": 0.6309664940413053, + "grad_norm": 1.5441378355026245, + "learning_rate": 0.00015793660697572848, + "loss": 0.1822, + "step": 40450 + }, + { + "epoch": 0.631122480813627, + "grad_norm": 1.3051613569259644, + "learning_rate": 0.00015792620785757367, + "loss": 0.3554, + "step": 40460 + }, + { + "epoch": 0.6312784675859487, + "grad_norm": 4.274336338043213, + "learning_rate": 0.00015791580873941892, + "loss": 0.2423, + "step": 40470 + }, + { + "epoch": 0.6314344543582704, + "grad_norm": 1.4855650663375854, + "learning_rate": 0.0001579054096212641, + "loss": 0.2158, + "step": 40480 + }, + { + "epoch": 0.6315904411305922, + "grad_norm": 0.26689472794532776, + "learning_rate": 0.00015789501050310936, + "loss": 0.3227, + "step": 40490 + }, + { + "epoch": 0.6317464279029138, + "grad_norm": 2.0525012016296387, + "learning_rate": 0.00015788461138495455, + "loss": 0.2475, + "step": 40500 + }, + { + "epoch": 0.6319024146752356, + "grad_norm": 1.1261099576950073, + "learning_rate": 0.0001578742122667998, + "loss": 0.2043, + "step": 40510 + }, + { + "epoch": 0.6320584014475572, + "grad_norm": 3.018596649169922, + "learning_rate": 0.000157863813148645, + "loss": 0.2309, + "step": 40520 + }, + { + "epoch": 0.632214388219879, + "grad_norm": 2.6361162662506104, + "learning_rate": 0.00015785341403049024, + "loss": 0.2583, + "step": 40530 + }, + { + "epoch": 0.6323703749922006, + "grad_norm": 1.5715960264205933, + "learning_rate": 0.00015784301491233543, + "loss": 0.2937, + "step": 40540 + }, + { + "epoch": 0.6325263617645224, + "grad_norm": 0.03241211175918579, + "learning_rate": 0.00015783261579418068, + "loss": 0.4566, + "step": 40550 + }, + { + "epoch": 0.6326823485368441, + "grad_norm": 0.5787049531936646, + "learning_rate": 0.00015782221667602587, + "loss": 0.0956, + "step": 40560 + }, + { + "epoch": 0.6328383353091658, + "grad_norm": 2.3547847270965576, + "learning_rate": 0.00015781181755787112, + "loss": 0.1877, + "step": 40570 + }, + { + "epoch": 0.6329943220814875, + "grad_norm": 2.2037112712860107, + "learning_rate": 0.0001578014184397163, + "loss": 0.0953, + "step": 40580 + }, + { + "epoch": 0.6331503088538092, + "grad_norm": 1.3263297080993652, + "learning_rate": 0.00015779101932156153, + "loss": 0.1661, + "step": 40590 + }, + { + "epoch": 0.6333062956261309, + "grad_norm": 0.20338000357151031, + "learning_rate": 0.00015778062020340675, + "loss": 0.1421, + "step": 40600 + }, + { + "epoch": 0.6334622823984526, + "grad_norm": 1.156614065170288, + "learning_rate": 0.00015777022108525197, + "loss": 0.1927, + "step": 40610 + }, + { + "epoch": 0.6336182691707744, + "grad_norm": 4.7294416427612305, + "learning_rate": 0.0001577598219670972, + "loss": 0.2359, + "step": 40620 + }, + { + "epoch": 0.633774255943096, + "grad_norm": 0.07647063583135605, + "learning_rate": 0.0001577494228489424, + "loss": 0.2377, + "step": 40630 + }, + { + "epoch": 0.6339302427154178, + "grad_norm": 1.5807030200958252, + "learning_rate": 0.00015773902373078763, + "loss": 0.1148, + "step": 40640 + }, + { + "epoch": 0.6340862294877394, + "grad_norm": 0.17653131484985352, + "learning_rate": 0.00015772862461263285, + "loss": 0.3267, + "step": 40650 + }, + { + "epoch": 0.6342422162600612, + "grad_norm": 2.8481879234313965, + "learning_rate": 0.00015771822549447807, + "loss": 0.2792, + "step": 40660 + }, + { + "epoch": 0.6343982030323828, + "grad_norm": 1.6328898668289185, + "learning_rate": 0.0001577078263763233, + "loss": 0.1816, + "step": 40670 + }, + { + "epoch": 0.6345541898047046, + "grad_norm": 1.9256733655929565, + "learning_rate": 0.0001576974272581685, + "loss": 0.5716, + "step": 40680 + }, + { + "epoch": 0.6347101765770262, + "grad_norm": 2.0054070949554443, + "learning_rate": 0.00015768702814001373, + "loss": 0.2331, + "step": 40690 + }, + { + "epoch": 0.634866163349348, + "grad_norm": 0.592220664024353, + "learning_rate": 0.00015767662902185895, + "loss": 0.1639, + "step": 40700 + }, + { + "epoch": 0.6350221501216697, + "grad_norm": 0.7444786429405212, + "learning_rate": 0.00015766622990370417, + "loss": 0.1833, + "step": 40710 + }, + { + "epoch": 0.6351781368939914, + "grad_norm": 1.4234321117401123, + "learning_rate": 0.00015765583078554938, + "loss": 0.4297, + "step": 40720 + }, + { + "epoch": 0.6353341236663131, + "grad_norm": 3.897197723388672, + "learning_rate": 0.0001576454316673946, + "loss": 0.2512, + "step": 40730 + }, + { + "epoch": 0.6354901104386348, + "grad_norm": 0.4277057647705078, + "learning_rate": 0.00015763503254923982, + "loss": 0.2606, + "step": 40740 + }, + { + "epoch": 0.6356460972109566, + "grad_norm": 3.6413636207580566, + "learning_rate": 0.00015762463343108504, + "loss": 0.1645, + "step": 40750 + }, + { + "epoch": 0.6358020839832782, + "grad_norm": 0.6908426284790039, + "learning_rate": 0.00015761423431293026, + "loss": 0.185, + "step": 40760 + }, + { + "epoch": 0.6359580707556, + "grad_norm": 0.7075207829475403, + "learning_rate": 0.00015760383519477548, + "loss": 0.2374, + "step": 40770 + }, + { + "epoch": 0.6361140575279216, + "grad_norm": 10.855485916137695, + "learning_rate": 0.0001575934360766207, + "loss": 0.2817, + "step": 40780 + }, + { + "epoch": 0.6362700443002434, + "grad_norm": 1.455926537513733, + "learning_rate": 0.00015758303695846592, + "loss": 0.1535, + "step": 40790 + }, + { + "epoch": 0.636426031072565, + "grad_norm": 0.469672292470932, + "learning_rate": 0.00015757263784031114, + "loss": 0.2253, + "step": 40800 + }, + { + "epoch": 0.6365820178448868, + "grad_norm": 0.19487737119197845, + "learning_rate": 0.00015756223872215636, + "loss": 0.1672, + "step": 40810 + }, + { + "epoch": 0.6367380046172084, + "grad_norm": 0.022243143990635872, + "learning_rate": 0.00015755183960400158, + "loss": 0.4319, + "step": 40820 + }, + { + "epoch": 0.6368939913895302, + "grad_norm": 0.6900194883346558, + "learning_rate": 0.0001575414404858468, + "loss": 0.12, + "step": 40830 + }, + { + "epoch": 0.6370499781618518, + "grad_norm": 1.100761890411377, + "learning_rate": 0.00015753104136769202, + "loss": 0.164, + "step": 40840 + }, + { + "epoch": 0.6372059649341736, + "grad_norm": 1.8163270950317383, + "learning_rate": 0.00015752064224953724, + "loss": 0.3855, + "step": 40850 + }, + { + "epoch": 0.6373619517064953, + "grad_norm": 0.14638766646385193, + "learning_rate": 0.00015751024313138246, + "loss": 0.0791, + "step": 40860 + }, + { + "epoch": 0.637517938478817, + "grad_norm": 0.419525682926178, + "learning_rate": 0.00015749984401322768, + "loss": 0.4335, + "step": 40870 + }, + { + "epoch": 0.6376739252511388, + "grad_norm": 1.365848183631897, + "learning_rate": 0.00015748944489507293, + "loss": 0.1774, + "step": 40880 + }, + { + "epoch": 0.6378299120234604, + "grad_norm": 0.020374717190861702, + "learning_rate": 0.00015747904577691812, + "loss": 0.1826, + "step": 40890 + }, + { + "epoch": 0.6379858987957822, + "grad_norm": 3.35870623588562, + "learning_rate": 0.00015746864665876337, + "loss": 0.1956, + "step": 40900 + }, + { + "epoch": 0.6381418855681038, + "grad_norm": 0.41389143466949463, + "learning_rate": 0.00015745824754060856, + "loss": 0.304, + "step": 40910 + }, + { + "epoch": 0.6382978723404256, + "grad_norm": 4.447183609008789, + "learning_rate": 0.0001574478484224538, + "loss": 0.2884, + "step": 40920 + }, + { + "epoch": 0.6384538591127472, + "grad_norm": 0.428251713514328, + "learning_rate": 0.000157437449304299, + "loss": 0.0913, + "step": 40930 + }, + { + "epoch": 0.638609845885069, + "grad_norm": 1.440173864364624, + "learning_rate": 0.00015742705018614424, + "loss": 0.3341, + "step": 40940 + }, + { + "epoch": 0.6387658326573906, + "grad_norm": 0.37606480717658997, + "learning_rate": 0.00015741665106798944, + "loss": 0.1545, + "step": 40950 + }, + { + "epoch": 0.6389218194297124, + "grad_norm": 0.08015301823616028, + "learning_rate": 0.00015740625194983468, + "loss": 0.2598, + "step": 40960 + }, + { + "epoch": 0.639077806202034, + "grad_norm": 1.5615071058273315, + "learning_rate": 0.00015739585283167988, + "loss": 0.1958, + "step": 40970 + }, + { + "epoch": 0.6392337929743558, + "grad_norm": 0.7406660914421082, + "learning_rate": 0.00015738545371352512, + "loss": 0.2646, + "step": 40980 + }, + { + "epoch": 0.6393897797466774, + "grad_norm": 1.9117989540100098, + "learning_rate": 0.00015737505459537032, + "loss": 0.2309, + "step": 40990 + }, + { + "epoch": 0.6395457665189992, + "grad_norm": 4.485423564910889, + "learning_rate": 0.00015736465547721556, + "loss": 0.1564, + "step": 41000 + }, + { + "epoch": 0.639701753291321, + "grad_norm": 2.9847829341888428, + "learning_rate": 0.00015735425635906075, + "loss": 0.1701, + "step": 41010 + }, + { + "epoch": 0.6398577400636426, + "grad_norm": 1.0845332145690918, + "learning_rate": 0.000157343857240906, + "loss": 0.1097, + "step": 41020 + }, + { + "epoch": 0.6400137268359644, + "grad_norm": 1.2486369609832764, + "learning_rate": 0.0001573334581227512, + "loss": 0.1147, + "step": 41030 + }, + { + "epoch": 0.640169713608286, + "grad_norm": 0.49971750378608704, + "learning_rate": 0.0001573230590045964, + "loss": 0.1771, + "step": 41040 + }, + { + "epoch": 0.6403257003806078, + "grad_norm": 8.242999076843262, + "learning_rate": 0.00015731265988644163, + "loss": 0.2553, + "step": 41050 + }, + { + "epoch": 0.6404816871529294, + "grad_norm": 0.251371830701828, + "learning_rate": 0.00015730226076828685, + "loss": 0.1483, + "step": 41060 + }, + { + "epoch": 0.6406376739252512, + "grad_norm": 0.4849778711795807, + "learning_rate": 0.00015729186165013207, + "loss": 0.3369, + "step": 41070 + }, + { + "epoch": 0.6407936606975728, + "grad_norm": 0.21417611837387085, + "learning_rate": 0.0001572814625319773, + "loss": 0.1097, + "step": 41080 + }, + { + "epoch": 0.6409496474698946, + "grad_norm": 0.6047383546829224, + "learning_rate": 0.0001572710634138225, + "loss": 0.2729, + "step": 41090 + }, + { + "epoch": 0.6411056342422162, + "grad_norm": 1.8839526176452637, + "learning_rate": 0.00015726066429566773, + "loss": 0.451, + "step": 41100 + }, + { + "epoch": 0.641261621014538, + "grad_norm": 1.1051620244979858, + "learning_rate": 0.00015725026517751295, + "loss": 0.1019, + "step": 41110 + }, + { + "epoch": 0.6414176077868596, + "grad_norm": 7.545604705810547, + "learning_rate": 0.00015723986605935817, + "loss": 0.5734, + "step": 41120 + }, + { + "epoch": 0.6415735945591814, + "grad_norm": 0.2164558321237564, + "learning_rate": 0.0001572294669412034, + "loss": 0.0858, + "step": 41130 + }, + { + "epoch": 0.641729581331503, + "grad_norm": 0.0324564203619957, + "learning_rate": 0.0001572190678230486, + "loss": 0.3937, + "step": 41140 + }, + { + "epoch": 0.6418855681038248, + "grad_norm": 0.1620205044746399, + "learning_rate": 0.00015720866870489383, + "loss": 0.1493, + "step": 41150 + }, + { + "epoch": 0.6420415548761464, + "grad_norm": 1.174554467201233, + "learning_rate": 0.00015719826958673905, + "loss": 0.2502, + "step": 41160 + }, + { + "epoch": 0.6421975416484682, + "grad_norm": 0.43777403235435486, + "learning_rate": 0.00015718787046858427, + "loss": 0.2157, + "step": 41170 + }, + { + "epoch": 0.64235352842079, + "grad_norm": 1.9883239269256592, + "learning_rate": 0.0001571774713504295, + "loss": 0.2866, + "step": 41180 + }, + { + "epoch": 0.6425095151931116, + "grad_norm": 0.18113389611244202, + "learning_rate": 0.0001571670722322747, + "loss": 0.2426, + "step": 41190 + }, + { + "epoch": 0.6426655019654334, + "grad_norm": 1.009268879890442, + "learning_rate": 0.00015715667311411993, + "loss": 0.1776, + "step": 41200 + }, + { + "epoch": 0.642821488737755, + "grad_norm": 1.7247790098190308, + "learning_rate": 0.00015714627399596515, + "loss": 0.3442, + "step": 41210 + }, + { + "epoch": 0.6429774755100768, + "grad_norm": 0.03508676588535309, + "learning_rate": 0.00015713587487781037, + "loss": 0.197, + "step": 41220 + }, + { + "epoch": 0.6431334622823984, + "grad_norm": 0.5332340598106384, + "learning_rate": 0.0001571254757596556, + "loss": 0.1786, + "step": 41230 + }, + { + "epoch": 0.6432894490547202, + "grad_norm": 0.3229464590549469, + "learning_rate": 0.0001571150766415008, + "loss": 0.1483, + "step": 41240 + }, + { + "epoch": 0.6434454358270418, + "grad_norm": 1.2618160247802734, + "learning_rate": 0.00015710467752334603, + "loss": 0.3074, + "step": 41250 + }, + { + "epoch": 0.6436014225993636, + "grad_norm": 2.8442862033843994, + "learning_rate": 0.00015709427840519125, + "loss": 0.2369, + "step": 41260 + }, + { + "epoch": 0.6437574093716852, + "grad_norm": 1.3382997512817383, + "learning_rate": 0.00015708387928703647, + "loss": 0.3876, + "step": 41270 + }, + { + "epoch": 0.643913396144007, + "grad_norm": 0.08462920784950256, + "learning_rate": 0.00015707348016888168, + "loss": 0.092, + "step": 41280 + }, + { + "epoch": 0.6440693829163286, + "grad_norm": 2.472012996673584, + "learning_rate": 0.0001570630810507269, + "loss": 0.2023, + "step": 41290 + }, + { + "epoch": 0.6442253696886504, + "grad_norm": 4.250405311584473, + "learning_rate": 0.00015705268193257212, + "loss": 0.1646, + "step": 41300 + }, + { + "epoch": 0.644381356460972, + "grad_norm": 0.7855823040008545, + "learning_rate": 0.00015704228281441734, + "loss": 0.3743, + "step": 41310 + }, + { + "epoch": 0.6445373432332938, + "grad_norm": 0.5818260908126831, + "learning_rate": 0.00015703188369626256, + "loss": 0.2424, + "step": 41320 + }, + { + "epoch": 0.6446933300056156, + "grad_norm": 1.2513214349746704, + "learning_rate": 0.00015702148457810778, + "loss": 0.145, + "step": 41330 + }, + { + "epoch": 0.6448493167779372, + "grad_norm": 0.02137545309960842, + "learning_rate": 0.000157011085459953, + "loss": 0.0825, + "step": 41340 + }, + { + "epoch": 0.645005303550259, + "grad_norm": 1.525722861289978, + "learning_rate": 0.00015700068634179822, + "loss": 0.1914, + "step": 41350 + }, + { + "epoch": 0.6451612903225806, + "grad_norm": 0.05738039314746857, + "learning_rate": 0.00015699028722364344, + "loss": 0.1357, + "step": 41360 + }, + { + "epoch": 0.6453172770949024, + "grad_norm": 1.2557497024536133, + "learning_rate": 0.00015697988810548866, + "loss": 0.1407, + "step": 41370 + }, + { + "epoch": 0.645473263867224, + "grad_norm": 1.3528168201446533, + "learning_rate": 0.00015696948898733388, + "loss": 0.3959, + "step": 41380 + }, + { + "epoch": 0.6456292506395458, + "grad_norm": 0.4063834547996521, + "learning_rate": 0.0001569590898691791, + "loss": 0.2874, + "step": 41390 + }, + { + "epoch": 0.6457852374118674, + "grad_norm": 2.3102779388427734, + "learning_rate": 0.00015694869075102432, + "loss": 0.1858, + "step": 41400 + }, + { + "epoch": 0.6459412241841892, + "grad_norm": 0.537541925907135, + "learning_rate": 0.00015693829163286954, + "loss": 0.2123, + "step": 41410 + }, + { + "epoch": 0.6460972109565108, + "grad_norm": 3.3351657390594482, + "learning_rate": 0.00015692789251471476, + "loss": 0.3811, + "step": 41420 + }, + { + "epoch": 0.6462531977288326, + "grad_norm": 1.269486427307129, + "learning_rate": 0.00015691749339655998, + "loss": 0.1698, + "step": 41430 + }, + { + "epoch": 0.6464091845011543, + "grad_norm": 0.900663435459137, + "learning_rate": 0.0001569070942784052, + "loss": 0.1036, + "step": 41440 + }, + { + "epoch": 0.646565171273476, + "grad_norm": 2.849454164505005, + "learning_rate": 0.00015689669516025042, + "loss": 0.2081, + "step": 41450 + }, + { + "epoch": 0.6467211580457977, + "grad_norm": 0.3679705858230591, + "learning_rate": 0.00015688629604209564, + "loss": 0.085, + "step": 41460 + }, + { + "epoch": 0.6468771448181194, + "grad_norm": 1.7488338947296143, + "learning_rate": 0.00015687589692394086, + "loss": 0.1924, + "step": 41470 + }, + { + "epoch": 0.6470331315904412, + "grad_norm": 4.973154544830322, + "learning_rate": 0.00015686549780578608, + "loss": 0.5048, + "step": 41480 + }, + { + "epoch": 0.6471891183627628, + "grad_norm": 0.4881376028060913, + "learning_rate": 0.0001568550986876313, + "loss": 0.1993, + "step": 41490 + }, + { + "epoch": 0.6473451051350846, + "grad_norm": 0.0297964196652174, + "learning_rate": 0.00015684469956947652, + "loss": 0.1568, + "step": 41500 + } + ], + "logging_steps": 10, + "max_steps": 192324, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2.527175245824e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/training_args.bin b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..51ec8c074aec62c737bdc86f3c7f1d33bd5739dc Binary files /dev/null and b/HVU_QA/t5-viet-qg-finetuned/checkpoint-41500/training_args.bin differ